text
stringlengths
144
682k
Hartree Fock method: A simple explanation Hartree Fock method If you are interested in learning computational chemistry, you cannot ignore the Hartree-Fock theory. Although researchers rarely use the HF method directly in their studies nowadays, it is the basis for almost all electronic structure methods. The HF method is the practical approach for solving quantum mechanics problems, and when it comes to describing the electronic structure of atoms and molecules, there is no substitute for quantum mechanics. However, comprehending the HF theory is a tough challenge, and its difficulty convinces many novice computational chemists to skip it. If truth be told, I had been doing so for some time, but then I realised that the theory would haunt me whenever I tried to learn a new topic in computational chemistry. Mastering the HF theory needs a solid understanding of linear algebra and a lot of intellectual effort. However, in this article, I tried my best to present a straightforward yet insightful explanation of the HF theory, which can be used as a starting point for mastering computational methods. How to read this article If you know nothing of the Hartree Fock method and just want a quick definition, you can only read the next section. But, If you are a computational researcher who wants a solid understanding of the HF theory and how computers solve electronic structure problems, read the entire essay sequentially. The last section describes how the HF method works, but you’ll need to know a few concepts from the former passages to grasp that section. Moreover, a moderate understanding of quantum mechanics is required. What is the Hartree Fock method? In a nutshell, the Hartree-Fock method is an approximate, iterative computational method for solving the time-independent Schrodinger equation of many-body electronic systems. A proper description of atomic systems is only possible by applying quantum mechanics rules and solving the legendary Schrodinger equation, which explains the motion of atomic and subatomic particles. But as we know from quantum mechanics textbooks, the Schrodinger equation is only solvable for a few elementary single-electron, hydrogen-like systems exactly. Accordingly, we need approximate methods for more complex systems. There are two fundamental theories to approximate the electronic structure of multi-electron systems in quantum chemistry: the valence bond theory and molecular orbital theory, which is equivalent to the Hartree-Fock theory. The HF approach is easier to code for computers and hence is the dominant approximate method for solving the time-independent Schrodinger equation. The HF approximation breaks down a multi-electron wavefunction into a set of one-electron wavefunctions, called molecular orbitals—indeed, the prevalent concept of molecular orbitals, which is in the mind of every chemist, originates from HF theory. We also call the HF method the self-consistent field method because of its particular iterative way of finding the Schrodinger equation solution, which I will explain in the next section. As I will explain later, the HF method itself has insufficient accuracy because it does not account for electron correlation. As a result, researchers rarely use it directly these days. But the HF model is the foundation theory behind most of the popular electronic structure methods. Indeed, we have many accurate computational methods that start with the HF equation and then fix its electron correlation somehow. They are referred to as post-Hartree Fock methods. Moreover, all of the semi-empirical methods and the Kohn–Sham implementation of DFT are based on the Hartree-Fock theory. Figure 1 shows the relation between HF and other electronic structure methods. The many-body electronic problem As I mentioned before, the main problem in theoretical chemistry is to solve the Schrodinger equation for a many-body electronic system consisting of nuclei and electrons. This section will briefly describe the problem and then explain the brilliant Hartree and Fock’s solution. Let’s start with the general form of the non-relativistic time-independent Schrodinger equation. Schrödinger equation equation 01 To solve this equation for any system, we must know the system’s Hamiltonian and wave function. According to quantum mechanics, the Hamiltonian operator for a many-body system composed of several nuclei and electrons is: molecular Hamiltonian equation 02 The above equation is in atomic units for simplicity. Its terms represent the kinetic energy of electrons, the kinetic energy of nuclei, the coulomb attraction between electrons and nuclei, repulsion between electrons and repulsion between nuclei, respectively. You can find the derivation of this equation in quantum chemistry textbooks. Nevertheless, the Born-Oppenheimer approximation states that, since nuclei are much heavier than electrons and move a lot slower, we can consider the electrons moving in the field of fixed nuclei. As a result, we can equate the second term in the above equation with zero and consider the last term as a constant, which can be neglected because adding any constant to an operator does not affect the operator eigenfunctions and only adds to operator eigenvalues. The remaining terms are known as the electronic Hamiltonians, and they represent the motion of N electrons in the field of M fixed nuclei. the electronic Hamiltonian equation 03 If we solve the Schrodinger equation for electronic Hamiltonian, the solution includes the electronic energy and the electronic wave function, which describes the motion of the electrons. electronic Schrodinger equation equation 04 The electronic wave function depends explicitly on the electronic coordinates but parametrically on the nuclear coordinates, which means that, although the electronic wave function does not include the nuclear coordinates, Ψelec is a different function for different arrangements of the nuclei. The total energy for fixed nuclei also consists of the constant nuclear repulsion. total molecular energy equation 05 If we can solve the electronic Shrodinger equation, we can describe the motion of the nuclei by introducing a nuclear Hamiltonian under the same assumptions used to derive the electronic equation. So the main problem in computational chemistry is to solve the electronic Shrodinger equation, which is the goal of the Hartree-Fock method. To solve the electronic Shrodinger equation, we have no choice but to break down the N electron equation into a set of one-electron equations, namely. A set of one-electron equations equation 06 To do so, The electronic Hamiltonian (equation 3) must be separable into a set of one-electron ones. saparate the electronic Hamiltonian into a set of one-electron ones equation 07 The first term in this equation is the sum of single-electron kinetic energies. The second is the sum of attractions between each electron and all nuclei. Therefore, the two first terms are separable—note that the separation is based on electrons. But, the last term is the sum of repulsion between all electron-pairs and is not divisible into single electron terms. Hartree suggested that we can approximate the electron-electron repulsion in an average way. This means, instead of calculating the repulsion for all electron pairs, we can calculate the repulsion between each electron and an average field of all other electrons. By doing so, we can separate the third term into a set of one-electron terms and write the Hamiltonian as a sum of one-electron operators: the Hamiltonian as a sum of one-electron operators equation 08 vHF is the average potential that the i-th electron experiences because of the other electrons, and f(x) is a one-electron operator called the Fock operator. Using the Fock operator, we can break down the electronic Shrodinger equation into a set of one-electron equations. Hartree Fock equation equation 09 Although this equation is a one-electron problem, the Hartree-Fock potential (vHF) depends on the entire system’s wave function. Therefore, this is a nonlinear problem and must be solved iteratively.  The HF method’s basic idea is that, using an initial trial wave function, the average field is calculated, then the eigenvalue equation (equation 9) is solved using this average field, and a new wave function is produced. The new wave function is used to calculate a new average field, and this procedure is repeated until some convergence criteria are satisfied. The self-consistent-field method (SCF) is another name for the HF approach, which I will explain in detail in the last section. Hartree Fock wave function and the Slater determinants Now that we got to know the Hamiltonian in equation 1, we must specify the form of the wave function for many-electron systems. First, the wave function must be separable to satisfy equation 4. Hartree approximately considered the electrons uncorrelated (independent from each other) to build a separable wave function. The square of the wavefunction is the probability of finding electrons in a specific volume of space, and the probability of occurring two independent events at the same time is equal to the product of their individual probabilities. Thus the electronic wavefunction of N uncorrelated electrons must be equal to the product of the one-electron wavefunctions. the Hartree Product equation 10 This equation is called the Hartree Product. The χ(x)s are one-electron wave functions referred to as spin-orbitals, and each of them is a function of three spatial coordinates and one quantum property called spin coordinate. However, the Hartree product has a severe flow as an electronic wave function. That is, it is not consistent with the indistinguishability of electrons. Note that if we interchange two electrons in the Hartree product, the result term is different. interchange two electrons in the Hartree product equation 11 Moreover, according to the Pauli exclusion principle, an electronic wave function must be antisymmetric, which means that with interchanging the space and spin coordinates of any two electrons, the sign of wave function alters. antisymmetric wavefunction equation 12 To satisfy these requirements, we can use certain linear combinations of Hartree products—note that if a wave function is acceptable, its linear combinations are also acceptable. Consider a two-electron wave function. If electron one and electron two occupy χi and χj spin-orbitals, respectively, we have Hartree product 1 equation 13 Also, we can make another Hartree product by putting electron one in χj and electron two in χi. Hartree product 2 equation 14 None of these Hartree products is antisymmetric, nor are their electrons indistinguishable, but we can make an antisymmetric wave function by combining them linearly. antisymmetric wave function equation 15 The 2 is a normalization factor, and the minus sign ensures that the wave function is antisymmetry. We can write this equation as a 2-by-2 determinant for the sake of tidiness. 2-by-2 Slatter determinant equation 16 Likewise, every N-electron wave function can be written as an N-by-N determinant called the Slatter determinant. Slatter determinant equation 17 Note that each row belongs to an electron in the Slater determinant, and each column represents a spin-orbital. The Hartree product is an entirely uncorrelated wave function, but the anti-symmetrised Slater determinant introduces some degree of correlation between electrons. Namely, the electrons with parallel spin can not be in the same spatial orbitals since, in that case, two rows of Slater determinant become equal—remember that if two rows of a determinant are identical, then the determinant is zero. Therefore the HF method takes into account a type of correlation, which is called exchange-correlation. However, because the HF method does not account for correlation between electrons with opposite spins, it usually is referred to as an uncorrelated method. The Slater determinant can be built by different combinations. Therefore, we can make multiple Slater determinants. Considering the two-electron wave function case, we have two spatial orbitals ψ1 and ψ2 and four spin orbitals constructed from them, ψ1α, ψ1β, ψ2α, and ψ2β. But in this case, to build a Slater determinant, we only need two spin orbitals, and there are six different possible combinations to make a two-electron Slater determinant out of four spin-orbitals when all permutations are taken into account. Six Slater determinants produced by four spin orbitals and corresponding states Figure 2. Four spin-orbitals produce six Slater determinants. A concise notation was used to show the Slatter determinants for the sake of tidiness. Generally, having N electrons and M spin-orbitals, we can construct $latex \displaystyle \frac{{M!}}{{N!(M-N)!}}$ different Slater determinants. However, the Hartree-Fock approach only uses one of them called the Hartree-Fock ground state determinant. The other determinants can represent the excited states of the electronic system. Also, a linear combination of these determinants is a more accurate wave function, which more precisely considers the correlation between electrons. A more elaborate and computationally cumbersome method, configuration interaction (CI), utilises the linear combination of these determinants to describe electronic systems. Using a single Slater determinant arguably equals treating electron-electron repulsion in an average way, which is the essence of the Hartree-Fock approximation. Basis sets and the Roothaan equations Now that we know that the electronic wave function can be built from the linear combination of spin-orbitals, the critical question is how to find spin-orbitals. Each spin-orbital consists of two parts, the spatial function and the spin function. The spatial function is a function of three electron spatial coordinates, and the spin function is related to electron spin, which is a quantum attribute. a spin-orbital equation 18 Since the Hamiltonian Operator (equation 8) does not affect the spin function, we can consider the spin function as a constant coefficient. spin function as a constant coefficient equation 19 Therefore to solve the problem, we only need to determine the spatial part of spin-orbitals. Clemens C. J. Roothaan and George G. Hall, both graduate students at the time, independently showed that the spatial component could be regarded as a set of known spatial basis functions, referred to as the basis set. basis set equation 20 C and φ represent expansion coefficients and known functions in this equation, respectively. The importance of the Roothan-Hall equation is that it reduces the problem of finding the HF molecular orbitals to the problem of calculating a set of expansion coefficients, which can be represented by a coefficient matrix. Therefore, we can write the Hartree Fock equation (equation 9) as a solvable matrix equation that can be proceeded by computer software. the Hartree Fock equation using basis sets equation 21 In terms of accuracy,  if we could use any complete basis set, which has an infinite set of K and spans the entire phase-space, the expansion would be exact regardless of the form of the basis functions. But, a finite basis set is only exact in the space spanned by its functions, so it is crucial which basis functions compose the basis set; however, as a rule of thumb, the greater the K and larger the basis set, the more precise are the molecular orbitals. But, using a huge basis set is not a practical option in real-world computational problems because the cost of electronic structure computations increases rapidly with the size of the basis set. Therefore we always have to compromise the basis set size. The Hartree Fock equation This section will study the HF equation and its components, although I will not provide a detailed derivation because it is meant to be an easy explanation. A keen reader can find a thorough explanation in this reference.. One and two-electron integrals As I said before, the Hartree-Fock method aims to solve the electronic Schrodinger equation (equation 4) for many-body electronic systems. electronic Schrodinger equation equation 04 This is an eigenvalue equation, and by solving the equation, I mean finding Ψelec vectors and Eelec values that fit the equation. Unfortunately, we can not solve this equation exactly for many-body electronic systems. However, we can find approximate solutions for it using the variation theorem. The variational principle states that any approximate wave function has an energy value above or equal to the exact energy. In other words, if we minimize the electronic energy with respect to the spin-orbitals, we can find the closest wave function to the exact one in the restriction of Hartree-Fock theory, which is called the Hartree-Fock ground state wave function. To apply the variation theorem, Let’s first rearrange equation 4 for energy. apply the variation theorem equation 22 We can substitute Helec with equation 3 and Selec with a single Slater determinant to find the energy equation in terms of Hartree-Fock approximation. the energy equation in terms of Hartree-Fock approximation equation 23 In this equation, I used a linear short-hand notation for the Slater determinant for the sake of conciseness. The electronic Hamiltonian consists of two types of operators. The first two terms are one-electron operators. These operators depend only on the position or momentum of one electron and do not consider the interaction with other electrons. As a two-electron operator, the last term is the sum of repulsion between all electron-pairs and thus involves the position of two electrons simultaneously. electronic energy in HF approximation equation 24 Let us indicate the one-electron operators with hone for simplicity. indicate the one-electron operators equation 25 We can consider equation 25 as a matrix equation. There are rules for evaluating such equations depending on the type of operators. Using these rules, you can find the expression of energy as  Hartree Fock energy equation 26 I am not mad enough to bring the derivation of this equation in this article; instead, I want to draw your attention to the components of the equation—a geek reader can find it in this reference. The first summation is over all spin-orbitals, and the inside integration is over the coordinates of a single electron (dx1). This single electron can be any electron in that certain spin-orbitals; nevertheless, we choose the dummy variable of the integral to be x1 for convenience. This part of the equation is referred to as one-electron integrals. The second summation is over all pairs of spin-orbitals, and the ½ coefficient is to avoid counting the pairs twice. The integrals in this part are over the coordinates of two electrons (dx1, dx2). Similarly, these two electrons can be any electron, and the integrals are known as two-electron integrals. Note that there are only N one-electron integrals in the electronic energy equation, but the number of two-electron integrals is proportional to N4. That’s why two-electron integrals always take a long time to calculate. Tow-electron integrals calculation in gamess us Figure 3: A part of GAMESS US software output which shows two-electron integrals calculation for the benzene molecule in HF/3-21g computational level, note that there are 1957454 basis functions for such a small molecule. Applying the variation theorem Now that we have an expression for electronic energy, we can use the variation theorem to solve the electronic Schrodinger equation. According to the variation theorem, the exact wave function always has a lower energy value than any approximate wave function, so the most accurate HF wave function is at the minimum with respect to changing spin-orbitals (the basis set coefficients). We call the energy of this wave function the Hartree-Fock ground state energy (E0). As a result, δE0 is zero, and we can write the following equation. equation 27 In this Equation, CC represents the complex conjugate of the left-hand side terms produced by the differentiation process. applying the variation theorem to Hartree-Fock approximation Figure 4: the variation theorem and the Hartree Fock approximation If all the basis-set coefficients were independent, we could use the above equation to minimise the energy directly; however, because spin-orbitals are normal, some basis-set coefficients are dependent. The normalisation requirement of spin-orbitals implies that. The normalisation requirement of spin-orbitals equation 28 And concerning equation 19 and equation 20 we can write: normalisation requirement of spin-orbitals 2 equation 29 And because spin and spatial functions also are orthogonal, this equation can be reduced to normalization of basis-set coefficients equation 30 The above equation shows that in each spin-orbitals, one of the basis set coefficients (Cvi) is dependent on other coefficients, and therefore the E0 is a constraint function of the Cvis. In mathematics, the standard method for finding the local minima of constraint functions is Lagrange’s method of undetermined multipliers. In this method, we introduce a new variable called Lagrange’s multiplier and define a new function (L) related to the multiplier and all other variables; then, we minimize L for all variables. Now let’s imply the Lagrange method to find δE0. apply the Lagrange method to derive Hartree-Fock equation equation 31 Note that we built the Lagrange function simply by adding a zero to the energy expression because the spin-orbitals are normalised. normalized spin-orbitals equation 32 So the minimum of L and E occurs at the same spot where δL is equal to zero minimize Lagrange function equation 33 The second term in this equation can be written as derive Hartree Fock equation by Lagrange method equation 34 Where CC indicates the complex conjugate of the left-hand side terms. Using this equation and equation 27 we can write δL as derive Hartree Fock equation by Lagrange method 2 equation 35 This equation can be factored to derive Hartree Fock equation by Lagrange method 3 equation 36 We can neglect the complex conjugate part because if A+A*=0, then A=0. derive Hartree Fock equation by Lagrange method 4 equation 37 derive Hartree Fock equation by Lagrange method 5 equation 38 Coulomb and exchange operators Now let’s define coulomb (J) and exchange (K ) operators as coulomb operator equation 39 exchange operators equation 40 The coulomb operator has a straightforward interpretation. The exact Coulomb interaction between two electrons in the atomic unit is r12-1. Likewise, in the Hartree-Fock approximation, electron one in χa experiences a one-electron potential from χb equal to $ \displaystyle \int{{d{{x}_{2}}{{{\left| {{{\chi }_{b}}\left( {{{x}_{2}}} \right)} \right|}}^{2}}r_{{12}}^{{-1}}}}$ and because in equation 38, the summation is over all spin-orbitals, each electron feels Coulomb repulsion from all other electrons. Note that each integral in the summation is independent of the others, which is why the HF approximation does not account for the Coulomb correlation between electrons. On the other hand, The exchange operator has no perceivable classical interpretation and results from exchange-correlation between electrons with parallel spins, which I previously explained. The existence of the exchange operator in the HF equation implies that the HF method is not entirely uncorrelated and consider the correlation between electrons with parallel spins. Using the Coulomb and exchange operators, we can shorten equation 38 as HF equation equation 41 The inside brackets term is the definition of Fock operator f(x). Therefore non-canonical Hartree-Fock equation equation 42  This is the Hartree-Fock equation. Note that we separated an unsolvable many-electron problem (equation 4) into N solvable —although approximately— one-electron problems. The appearance of summation in the right-hand side of the equation is because the Slater determinant is formed from a set of spin-orbitals, which provide a certain degree of flexibility; namely, there can be different  spin-orbit combinations without changing the expectation value (E0). However, because ε is a Hermitian matrix, we can always find a unitary matrix that transforms ε into a diagonal matrix. The unique set of spin-orbitals corresponding to this specific diagonal ε matrix is called the set of canonical spin-orbitals.  It is typical to write equation 42 in its canonical (standard) form, the same as equation 9. Hartree Fock equation equation 9 The Self-Consistent field method; How computers solve the HF equation Now that we are familiar with the HF equation and how to build a proper wave function for it, In this final section, I can explain how computers solve the HF equation for many-body electronic systems. The specific procedure for solving the HF equation is called the self-consistent-field (SCF) method. SCF is another name for the HF method, and the two terms are used interchangeably in computational literature. Let’s look at the HF equation in its canonical form once again. Hartree Fock equation equation 9 χ represents a spin-orbital consisting of a spatial and a spin function in this equation. The spin function has only two values (α and β), and its value can be determined according to Pauli’s exclusion principle before SCF calculation. In the Hartree Fock approach, there are three formalisms to deal with spin function: restricted closed-shell, unrestricted open-shell and restricted open-shell. We can apply one of them according to the multiplicity of the electronic system and eliminate the spin function from equation 9. restricted closed-shell Hartree Fock equation equation 43 Thanks to Roothaan contribution to the SCF approach, we can consider the spatial function a set of known basis functions. spatial function as a set of known basis functions equation 44 By substituting the basis-set into equation 43, we can write the HF equation as the HF equation equation 45 This is an integro-differential equation; however, multiplying by Φ* on the left and integrating it, we can turn it into a matrix equation that is far more computer-friendly. Hartree Fock matrix equation equation 46 If you can not see these terms as a matrix equation, consider μ and v as matrix indices with values ranging from 1 to K. Now let’s define two new matrices.  1. The overlap matrix S with elements according to overlap matrix equation 47 The overlap matrix quantifies the degree of overlap between basis functions that are not orthogonal to each other in general. The diagonal elements of S are unity, and the off-diagonal elements have values ranging from one to zero. The sign of the off-diagonal elements depends on the relative sign of the two basis functions and their relative orientation. The more an off-diagonal element (S) approaches unity, the more its two basis functions overlap. 2. The Fock matrix F with elements according to the definition of the Fock matrix equation 48 The Fock matrix (F) is the matrix representation of the Fock operator (f) —a one-electron operator as I defined in the previous section— with the set of basis functions {Φ}. Because the summation indices (𝜇 and v) in equation 44 range from 1 to K, both the overlapped and Fock matrices are K×K square matrices. They are also Hermitian, which means we can orthogonalize them by a unitary matrix. With these definitions, now we can write the equation 46 as Roothann equation equation 49 Or more compactly as Roothann equation compact equation 50 In this equation, referred to as the Roothann equation, C is a K×K square matrix containing expansion coefficients of the basis functions (equation 44), and ε is a diagonal matrix of the orbital energies. basis set expansion coefficients equation 51 diagonal matrix of the orbital energies equation 52 Solving the Roothaan equation—or equivalently the Hartree-Fock equation—from the perspective of computational software means finding a proper coefficient matrix (C) that minimises the system’s total energy. The Roothaan equation (equation 50) is not a standard eigenvalue problem due to the non-orthogonality of the basis functions, which gives rise to the overlap matrix. Therefore, we need to consider a procedure to orthogonalise the basis functions. If we somehow manage to orthogonalise the basis functions, the overlap matrix becomes the unit matrix, and Roothaan’s equations turn into a set of standard eigenvalue problems. oothaan’s equations turn into a set of standard eigenvalue problems equation 53 If we have a set of non-orthogonal basis functions {Φ}, it is always possible to find a transformation matrix (X) that transforms the non-orthogonal set to an orthogonal one. transfer into orthogonal basis functions equation 54 orthogonal basis functions equation 55 We can convert the overlap matrix to a unit matrix using this transformation matrix. convert the overlap matrix to a unit matrix using this transformation matrix equation 56 There are two methods for finding the transformation matrix—symmetric orthogonalization and canonical orthogonalization—which I will not explain in this article. Nevertheless, having the X, we can orthogonalize the basis functions and eliminate the overlap matrix (S) from Roothaan’s equations. Afterwards, we can solve the equation directly by diagonalizing the Fock matrix. However, we have to transform all the two-electron integrals, which is very time-consuming. The SCF method uses a more efficient way to deal with the problem. Consider a new coefficient matrix (C’) related to the matrix by X a new coefficient matrix equation 57 Where X-1 is the inverse of the transformation matrix. Substituting C from the above equation into the Rootaan equations (equation 50) gives Substituting C into Rootaan equations equation 58 Multiplying both sides of this equation by the adjoint of the transformation matrix (X) gives equation 59 If we define a transformed Fock matrix (F’) such that transformed Fock matrix equation 60 We can define transformed Roothaan equations as transformed Roothaan equations equation 61 The SCF style of solving Roothan equations is finding the transformation matrix, transforming the Fock matrix and solving the transformed Roothaan equations. Finally, the coefficient matrix is calculated from equation 57. Indeed, the SCF instruction is more elaborate than this. The Roothaan equations are nonlinear and need to be solved by an iterative approach because the Fock matrix depends on the expansion coefficients. Before explaining the SCF procedure, we need to know the density matrix concept and the Fock matrix’s explicit form. The density matrix We know from quantum mechanics that the probability of finding an electron in a specific volume (dr), the charge density of that electron, is equal to |ψa(r)|2 dr. Now consider a closed-shell system described by a single determinant wave function, in which all occupied molecular orbitals contains two electrons. In this system, the total charge density is total charge density equation 62 Let us expand the molecular orbitals as a set of basis functions from equation 44 expand charge density as a set of basis functions equation 63 From this equation, we can define the density matrix as density matrix equation 64 The Density matrix is directly related to the expansion coefficients, and it can represent the Hartree-Fock wavefunction completely. The density matrix is important because, in the SCF instruction, the Fock operator is written in terms of the density matrix (you will see it in the following subsection), and it solves the HF equation iteratively concerning the charge density. The Hartree Fock codes first guess a charge density describing the position of the system’s electrons. Then they use this conjectural charge density to estimate an initial Fock operator. Using this operator, they solve one-electron Schrodinger-like equations (equation 43) to find the molecular orbitals i}. Then they can build a better charge density from these recent molecular orbitals and guess a more accurate Fock operator. The SCF algorithm repeats this procedure until the Fock operator no longer changes. In other words, the SCF procedure continues the loop of making the Fock operator and charge density from each other until the field (the two-electron part of the Fock operator) that produces a specific charge density becomes identical (consistent) with the field which would be calculated from that charge density. This is why we also call the HF method the self-consistent-field method. In the following subsection, we scrutinize the Fock operator to find an explicit expression for it which we will use in the final part to demonstrate how the Hartree Fock method works. The Fock matrix Let us have another look at equation 41. In this equation, the Fock operator is Fock operator equation 65 However, if we consider the simplest form of the Hartree Fock equation, the restricted closed-shell HF equation, the Fock operator slightly differs from the above equation (for proof, see section 3.4.1 of reference 1). the Fock operator for restricted closed-shell systems equation 66 Coulomb operator equation 67 exchange operator equation 68 For our purpose, it is easier to define a new operator, P12, which interchange electrons 1 and 2 and rewrite the exchange operator. permutation operator equation 69 equation 70 Now we can write the Fock operator as the Fock operator equation 71 And then incorporate the basis functions (equation 44) into it. incorporate the basis functions in Fock operator equation 72 We can pull expansion coefficients out of the integral and define them as a density matrix (equation 64). density matrix in the Fock operator equation 73 density matrix in the Fock operator 2 equation 74 Finally, using the definition of the Fock matrix (equation 48)  we can write Fock matrix equation 75 The first integral term, called core-Hamiltonian (Hcore), involves only one-electron operator (hone see the above passage) and is over the coordinates of one electron (r1) each time. The core-Hamiltonian represents an electron’s kinetic energy and nuclear attraction, neglecting the repulsion with other electrons. Unlike the full Fock matrix, the core-Hamiltonian matrix needs to be assessed only once as the beginning of the SCF procedure. The remaining part of the Fock matrix is a two-electron part, denoted by G, which consists of the density matrix and the two-electron integrals. Fock operator consists of the density matrix and the two-electron integrals equation 76 Fock operator one electorn and two-electron part equation 77 At this point, we have enough materials to properly describe the SCF procedure, which is the topic of the final part. The self-consistent field algorithm This section describes the actual procedure that computational codes perform to find the restricted closed-shell Hartree-Fock equation for molecules. The instruction for open-shell systems are pretty similar, and they are at the heart of many computational methods. The following passage is helpful for users and developers of the electronic structure software to understand what is happening under the shell of computational software. However, the prerequisite for this section is to understand the concepts and theories I have presented in the previous parts of this article, so if you find the following paragraph ambiguous, please go back and read the previous sections.   Note: the terms self-consistent-field and Hartree-Fock refer to the same method, so feel free to use them interchangeably.   The SCF method goes like this: 1. Specify the nuclear coordinates, atomic numbers, number of electrons, and the basis set from the input file. 2. Calculate the overlap matrix (S), The core-Hamiltonian (Hcore), and the first estimate of two-electron integrals.  3. Find the transformation matrix (X) by diagonalising the overlap matrix (S). 4. Guess the initial density matrix (D). 5. Calculate the two-electron matrix G (equation 76 and 77) from the density matrix (D) and the two-electron integrals. 6. Calculate the first Fock matrix (F) by adding H to G (equation 77). 7. Transform the Fock matrix using the transformation matrix, F′=XFX. 8. Solve the transformed Roothaan equations, F′C′=C′𝜺. 9. Calculate expansion coefficients, C=XC′. 10. From a new density matrix (D) using C and equation 64. 11. Compare the new density matrix with the old one and check how different they are. If the difference is greater than a specific convergence criterion, return to step 5. If the difference is smaller than the criterion, go to the final step. 12.  Calculate the total energy from the Fock and Density matrices. algorithm for self-consistent field theory Figure 5: the algorithm of self-consistent field (SCF) theory Leave a Comment Scroll to Top
Article plus Helping Material:Just click here to see.(1.5M, pdf). the full total focus (6). Plasma membranes are complicated fluids, containing an array of different parts. Approximately 50% of the?plasma membrane includes lipids by pounds, and 40% from the lipids, on the mole basis, are cholesterol substances (7, 8). Cholesterol can be loaded in plasma membranes, therefore relationships between cholesterol and additional constituentseither lipids or proteinscan be considered a main controller of mobile processes. Outcomes of cholesterol relationships have already been looked into for both cell and model membranes (9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27). The analysis of cholesterol relationships with additional lipids and with proteins garnered improved attention using the explicit formulation of the idea of rafts (28, 29, 30). The focus of cholesterol sequestered within a site would, necessarily, vary by area inside the membrane. The theory that the way of measuring concentration can be insufficient to reliably quantify the part of cholesterol in natural processes continues to be well documented in many ways. For example, it’s been proven that not absolutely all cholesterol can be free to connect to additional molecules by calculating its availability to externally added cholesterol oxidase (11, 20, 31, 32, 33, 34), poisons (35, 36, 37), or cyclodextrins (38, 39). The assays found in these and additional research of cholesterol relationships have clearly founded that focus and activity of cholesterol in plasma membranes isn’t the same (2). Although these assays possess definitively Filixic acid ABA demonstrated that focus of cholesterol Filixic acid ABA can be insufficient to determine its activity, they don’t give a universal thermodynamic scale to quantify the cholesterol chemical chemical or potential activity. A quantitative size is required, for instance, if chemical substance activity of cholesterol in various cells types or under assorted conditions should be likened. We therefore created a strategy to quantify the chemical substance potential of plasma membrane cholesterol. We denote this amount as (41). To look for the cholesterol focus in the aqueous MBCD remedy, we adapted the typical procedure predicated on oxidation Filixic acid ABA of cholesterol by cholesterol oxidase to create H2O2, which can be detected by responding the peroxide with 10-acetyl-3,7- dihydroxyphenoxazine to produce resorufin. Explicitly, 25 for Filixic acid ABA 15?min to eliminate any possible particles, as well as the filtrate was collected. Cholesterol content material in the filtrate was established the same manner for nucleated cells and changed into cholesterol chemical substance potential (discover below). The lipids from the RBCs had been extracted using the Bligh and Dyer technique (43), as well as the cholesterol content material of the extract was quantified. Identifying cholesterol chemical substance potential within plasma membranes of nucleated cells Cells were cultivated and seeded in 24-very well plates. Just before chemical substance potential experiments, press was aspirated and cells had been washed once, a brand new Tyrodes remedy was added, as well as the cells had been incubated for 30?min in 37C with plates on the rocking system. The bathing remedy of every well was changed by Tyrodes solutions including MBCD/cholesterol at different predetermined cholesterol chemical substance potentials (thought as preliminary chemical substance potential) that spanned the cholesterol saturation selection of 21C79%. Plates had been sealed to avoid evaporation and incubated for 15?min in 37C for the rocking system. Solutions from each well had been gathered and filtered using AcroPrep Progress Omega 30 MWCO plates (Pall) to eliminate particulate materials. This remedy was assayed for cholesterol, yielding the ultimate chemical substance potential. Although unlikely highly, if a part of cholesterol in MBCD had not been available to cholesterol oxidase, the result for the established Rabbit Polyclonal to MARK2 ln can be increasingly more adverse as cholesterol focus can be reduced below its saturation limit. Cholesterol focus in MBCD can be a non-linear, saturating function of cholesterol activity We mixed an Filixic acid ABA MBCD remedy preloaded with cholesterol and cholesterol within an organic remedy (squalane or hexadecane), allowed cholesterol to equilibrate, and established the cholesterol focus in the aqueous stage. Curves to calibrate the quantity of cholesterol destined to?MBCD were generated for each and every focus of MBCD we used (always 3?mg/mL MBCD in this specific article, except for is definitely a. Background: There is a dramatic rise in liver organ failure and numerous patients undergo liver organ transplant for life-saving reasons each year. of liver-specific genes, tyrosine cholesterol and aminotransferase 7 alpha-hydroxylase, were even more significant in Rabbit Polyclonal to Cytochrome P450 4F3 the feminine treated group weighed against the male treated group on day time 7 (p<0.05); however, after 30 days, there were no significant variations. Furthermore, hematoxylin and eosin and periodic acid-Schiff staining of liver sections shown the considerable liver regeneration post cell therapy in both organizations. Notably, data has shown that MenSCs could engraft into hurt liver tissues and result in the same effect in the regeneration of liver function in both genders. Summary: Results of this study introduce MenSCs therapy as a good alternative approach for liver fixing and regeneration which has no gender constraints. and studies which statement the potential of differentiation of MenSCs into hepatocyte-like cells and repair of AHF with transplantation of these stem cells in mice model 11,21, but you will find Cucurbitacin IIb no data about the outcome of Cucurbitacin IIb MenSCs transplantation into a male recipient. The most remarkable element about the MenSCs is definitely that they are gender-specific cells with several general variations with male cells which must be acknowledged prior to transplantation of them into the male Cucurbitacin IIb recipients. In addition, a greater understanding of the sex effect on varied stem cell populations is required to improve their greatest clinical efficacy. Consequently, in this study, an attempt was made to compare the effectiveness of MenSCs transplantation within the repair of AHF in male and female immunocompetent Balb/C mice with acute liver failure. The recipient’s gender was considered as a key point in predicting the effect of MenSCs transplantation on reducing the effects of toxic liver agents. The restorative effect of MenSCs was assessed by evaluation of the biochemical, histopathological, and molecular hepatic factors in both genders at different times. Materials and Methods Isolation and tradition of MenSCs Donors for menstrual blood were selected from healthy females aged 20C35 years. All donors authorized an informed consent authorized by the medical ethics committee of Avicenna Study Institute. About 5 of menstrual blood was collected using a Diva cup (Di-va International Co., Lunette, Finland) during the 1st 2 days of the menstrual cycle. The material of Diva cup were collected inside a collection tube comprising 2.5 fungizone (Gibco, UK), 100 streptomycin, 100 penicillin (Sigma-Aldrich, St. Louis, MO, USA), and 0.5 EDTA in Phosphate Buffered Saline (PBS). MenSCs were isolated and cultured from menstrual blood by the method explained previously 22. Briefly, menstrual blood mononuclear cells were separated by Ficoll-Hypaque (GE-Healthcare, Uppsala, Sweden) denseness gradient. After washing with PBS, the cell pellet was consequently cultured inside a T75 flask comprising complete Dulbeccos revised Eagles medium-F12 (DMEM-F12; Sigma-Aldrich) supplemented with 10% FBS, 2 L-glutamine, 100NEAA, 100 penicillin, 100 streptomycin, and 25 fungizone and taken care of at 37in a 5% humidified CO2 incubator for approximately 2 days. Following initial incubation, the non-adherent cells were washed away, leaving behind the adherent cell human population. 4C7 days later on, the 1st colonies appeared. After achieving 70C80% confluence, adherent cells were passaged using Trypsin/EDTA (Gibco). Immunophenotyping of MenSCs The cultured MenSCs were recognized via the manifestation of mesenchymal stem cell markers such as CD73, CD105, CD146, CD10, CD29, hematopoietic stem cell markers such CD34 and CD133, and embryonic stem cell marker including OCT4, SSEA-4 using a flow cytometric analysis, as. The hepatitis B computer virus (HBV) can’t be taken out completely from contaminated hepatocytes, due to the current presence of intrahepatic covalently shut round DNA (cccDNA). the clinical applications of M2BPGi and HBcrAg in CHB patients. Additionally, because brand-new potential therapeutic realtors that remove intrahepatic cccDNA are getting developed, monitoring of M2BPGi or HBcrAg may be ideal for evaluating therapeutic results as well as the clinical final results. To conclude, these will be suitable surrogate markers for predicting disease development. agglutinin-positive Macintosh-2 binding proteins 1. Launch Viral hepatitis can be an infectious disease and a respected global killer [1]. The hepatitis B trojan (HBV) causes severe and chronic an infection, frequently leading to liver-related illness and accounting for over 600, 000 liver-related deaths every year [2]. The majority of fresh HBV infections happen in highly endemic areas, such as China, Southeast Asia, and sub-Saharan Africa [3]. Chronic Cbll1 swelling during active HBV illness is definitely associated with continuing hepatocellular damage and cells restoration [4]. Consequently, liver fibrosis develops, having a progressive loss of liver function and improved risk of hepatocellular carcinoma (HCC). Up to 10% of chronic hepatitis B (CHB) individuals may progress to severe fibrosis and cirrhosis, a major risk element for HCC development [3]. Current HBV treatment ARS-1323 primarily seeks to prevent complications associated with progressive swelling and fibrosis, i.e., liver failure, decompensated liver cirrhosis, and HCC [5,6]. Regrettably, although nucleos(t)ide analogs (NAs) or interferons (IFNs) can efficiently suppress HBV replication, these treatments are not curative [7]. This is because these medicines do not directly target the key molecule responsible for intrahepatic viral persistence, covalently closed circular DNA (cccDNA). cccDNA is definitely a stable, extra-chromosomal transcriptional template for those HBV messenger RNAs (mRNAs) such as pregenomic RNA [8,9,10]. The amount and transcriptional activity of cccDNA in the hepatocytes are important for HBV disease progression and medical results [11]. To accomplish effective medical management of CHB, accurate laboratory data for analysis, treatment, monitoring, and prognostic assessment are essential (Number 1). Liver biopsy, regarded as the gold standard for assessment of the degree of liver disease, is definitely invasive and dangerous [12 possibly,13]. Furthermore, it really is susceptible to sampling mistakes and subjective result interpretation [14,15]. HCC medical diagnosis and security derive from the recognition of tumor markers mainly, such as for example alpha-fetoprotein (AFP) and proteins induced by supplement K lack or antagonist-II (PIVKA-II), and imaging methods [5,6,16]. There continues to be the necessity to get more reliable, noninvasive, and cost-effective biomarkers for CHB administration. Open in another window Amount 1 Chronic hepatitis B trojan ARS-1323 (HBV) infection linked to liver organ disease progression. The schematic shows the medical stages involved in the natural history of chronic hepatitis B (CHB). The novel serum biomarkers hepatitis B core-related antigen (HBcrAg) and ARS-1323 Mac pc-2 binding protein glycosylation isomer (M2BPGi) provide important prognostic data for effective management of CHB. It is important to monitor the individuals at high risk and to treat them in order to prevent liver complications, cirrhosis, and hepatocellular carcinoma (HCC) development. Abbreviations: HBV, hepatitis B disease; HCC, hepatocellular carcinoma; TACE, transcatheter arterial chemoembolization; TKIs, tyrosine kinase inhibitors; AFP, alpha fetoprotein; PIVKA-II, protein induced by vitamin K absence or antagonist-II; HBcrAg, hepatitis B core-related antigen; M2BPGi, Mac pc-2 binding protein glycan isomer; HBsAg, hepatitis B surface antigen. With this review, we evaluate two novel biomarkers ARS-1323 showing great potential for HBV analysis and prognostic evaluation. The first is hepatitis B core-related antigen (HBcrAg), a surrogate marker of intrahepatic HBV replication that has shown good correlation with standard HBV markers, such as HBV DNA and hepatitis B surface antigen (HBsAg) [17,18]. The second is Mac-2-binding proteins glycosylation isomer (M2BPGi), which really is a liver fibrosis marker with predicting HCC development [19]. We concentrate on the scientific utility of the markers as predictors of HBV-related HCC advancement. 2. HBV Normal Biomarkers and Background 2.1. HBV Replication Routine Hepatitis B virions (Dane contaminants) gain entrance in to the hepatocytes by binding towards the receptor sodium taurocholate co-transporting polypeptide (NTCP) [20] and potential hepatocyte-specific co-receptors over the cell surface area. The HBV envelope fuses using the membrane from the hepatocyte as well as the virion is normally endocytosed, launching the viral DNA (partly double-stranded round DNA), enclosed using the primary particle, in to the cytoplasm [21]. The viral envelope is normally dropped (uncoating). The viral nucleocapsid filled with the genomic DNA within a tranquil circular form is normally transported in to the nucleus. In the nucleus, the viral DNA polymerase synthesizes double-stranded DNA completely, which is normally changed into [21 cccDNA,22]. The forming of this cccDNA continues to be poorly recognized but is most likely via a DNA restoration mechanism [22] (restoration and cccDNA formation). The cccDNA is definitely then transcribed into the pregenomic and subgenomic mRNAs from the sponsor RNA polymerase II [21,22]. In response to tension cells must reprogram gene expression to adjust and survive quickly. stress-induced mRNP granule composition in different cell stress and types conditions. However, immediate measurements from the phenotypic and practical consequences of tension granule and control body formation lack. This leaves our knowledge of their jobs during tension incomplete. Continued study into the function of these granules will be an important part in elucidating how cells respond to and survive stressful environmental changes. Graphical Abstract Diverse macromolecular interactions lead to the phase separation of protein and RNA during stress. While the identities of many proteins and RNAs contained in these granules (tan spheres) have been elucidated recently, the function of this conserved EFNA2 compartmentalization of the cytoplasm during stress response is still an open question. Introduction Cells are frequently exposed to fluctuating, potentially adverse environmental conditions. To survive adverse changes they must rapidly alter gene expression in order to maintain internal homeostasis. The cellular reprogramming that occurs in response to a disruptive or inimical external fluctuation is broadly termed as stress response. Cellular stress response typically includes slowing or ceasing growth that is concomitant with repression of overall translation, though certain genes important for survival and repair are highly induced. Concurrently, while overall translation is repressed, many post-transcriptional regulatory proteins and mRNAs undergo a process called phase separation that results in the formation of concentrated, non-membranous cytoplasmic structures generally described as granules or foci. During stress, this phase separation process might segregate proteins and mRNAs in a way that is functionally important for the cell and that promotes survival. Therefore, these structures are a subject of emergent interest. Although much progress has been made recently to identify the proteins and mRNAs that reside in these granules and the physical characteristics that underlie their formation, there is little known about the phenotypic or functional consequences of their formation during stress and therefore how significantly they contribute to stress response. There are many different types of cellular granules involved in a wide variety of biological processes such as nucleoli, paraspeckles, PML bodies, and Cajal bodies in the nucleus as well as the stress-induced control physiques (PBs) and tension granules (SGs) in the cytoplasm. Right here we high light the cytoplasmic SGs and PBs, two well-studied mRNP granules that can be found across eukaryotes throughout a variety of difficult conditions such as for example exposure to temperature shock, oxidative tension, UV Pentagastrin irradiation, osmotic tension, and nutrient hunger. The forming of these mRNP granules, which happens on the size of mins after contact with tension stimuli, can be mediated with a physical procedure called liquid-liquid stage Pentagastrin separation (LLPS; discover Box 1 to find out more). There are normal biophysical characteristics plus some shared components between PBs and SGs aswell mainly because granule-specific features. It ought to be mentioned that, as the called SGs are broadly induced during tension aptly, PBs certainly are a little more organismal particular. induces noticeable PBs during tension response while mainly, in mammalian cells, little, microscopically noticeable PBs are constitutive however they become much bigger and even more abundant during tension. It will also be mentioned that most study into these stress-induced granules is conducted with candida and mammalian cell culture systems. Ultimately, we posit SGs Pentagastrin and PBs should be considered as distinct yet closely related mRNP granules; their properties and role in post-transcriptional gene expression during stress response is the focus of this review and we’ll address them independently, as PBs or SGs, and more generally together, as stress-induced mRNP granules, when suitable. Container 1: Liquid-liquid stage separation The parting of elements in a combination is certainly termed demixing. The demixing of fluids may take place if the power of relationship between macromolecules is certainly higher than the entropic energy decrease that comes from their homogeneous blending. This technique drives liquid-liquid stage separation of the molecular blend into two stages: a smaller sized level of a focused, condensed stage and a. Supplementary MaterialsS1 Fig: Host cell viability following infection with BCG-lux at MOI 5 and MOI 10. 14 days with a luminometric assay. Data are indicated as Replication Index, determined as the percentage between the Comparative Luminescence Products (RLU) acquired at your day indicated in shape as well as the RLU worth obtained at the start of tradition (t0). Data are representative of 2 independent cultures.(TIF) pone.0213150.s002.tif (175K) GUID:?AB404FCE-C708-4518-9566-BAD180A8CFDA S3 Fig: extract does not exert any direct toxic effects on host cells. DC (A, B), M1 (C, D) and M2 (E, F) (2×105/well) were stimulated with HyE-Ov at different concentration (1, 3, 9 and 27 mg/ml of equivalent plant material) for 3 days. Cells and supernatants were subjected to MTT Assay (A, C, E) and to CytoTox 96 Assay (B, D, F). Data are expressed as means SD of % of cell viability or % of cytotoxicity of triplicate cultures and are representative of 2 independent experiments performed on cells from different donors. *L. ssp. (HyE-Ov), a plant traditionally known for its anti-oxidative properties. The effects of HyE-Ov were tested on human monocyte derived dendritic cells (DC), type-1 (M1) and type-2 macrophages (M2) infected with Bacille Calmette-Gurin (BCG), used as a model of persistent intracellular bacterium. DC, M1 and M2 treated with HyE-Ov significantly enhanced their mycobactericidal activity, which was associated with phagosomal acidification in M1 and M2 and increase of phagosomal, but not mitochondrial ROS production in M1, M2, and DC. Treatment of BCG-infected DC with HyE-Ov significantly reduced TNF- and IL-12 production and increased TGF- synthesis. Finally, experiments were repeated using eight different HPLC fractions of HyE-Ov. Results showed that the capability to activate anti-microbial and anti-inflammatory response is usually shared by different fractions, suggesting that diverse bioactive molecules are present within the hydroalcoholic extract. Benzocaine hydrochloride Altogether, these results show that HyE-Ov promotes anti-mycobacterial innate immunity and limits inflammatory response and suggest that this herb extract may be exploitable as phytocomplex or nutraceutical for novel host-directed therapeutic approaches. Introduction Since ancient times, man has used plants to obtain therapeutic benefits and there is now an increasing Benzocaine hydrochloride scientific interest for their biological Rabbit polyclonal to MMP1 properties as they can provide a plethora of novel molecules of pharmacological interest [1]. An assessment of all Food and Drug Administration (FDA)- and European Medicine Agency (EMA)-approved molecules reveals that natural products and their derivatives represent over one-third of all new molecules, with one quarter of them derived from plants [2] as they may contain a variety of biologically active secondary metabolites with possible therapeutic value against infectious as well as noninfectious diseases and cancer [1, 3C5]. Among plants of potential medical interest, L. is known for its properties as expectorant, antimicrobial and carminative [6]. genus belongs to the Lamiaceae family and most of its species are distributed around the Mediterranean area, Eurasia and the North of Africa, where it is used in traditional medication for the treating cold, cough, respiratory and digestion disorders [6, 7]. The healing aftereffect of oregano in traditional medication was related to the antimicrobial generally, antioxidant and anti-inflammatory properties shown by it is phenolic substances [8C10]. This natural impact is certainly taken care of in ethanol and methanol macerations, infusion, decoction, hydroalcoholic remove, apart from in essential natural oils [11C16]. Specifically, anti-inflammatory activity of is because of the ability of its ingredients to reduce creation of nitric oxide, to diminish and raise the creation of inflammatory and anti-inflammatory cytokines, [17C20] respectively. In addition, a primary antimicrobial activity continues to be reported against fungi [21, 22], gram-negative and gram-positive bacterias [13, 14, 23]. This impact was induced with the phenolic monoterpenes, carvacrol and thymol, which as well as their precursors (MTB) but still represents one of many causes of loss of life by one pathogen Benzocaine hydrochloride infection, world-wide. Based on the last record by World Wellness Firm (WHO), in 2017, 10.0 million people fell with TB and 1 ill.6 an incredible number of them passed away [25]. Today, the introduction of mycobacterial strains, pathogenic to human beings, endowed with level of resistance to first-line (Multi-Drug Resistant, MDR) and second-line antibiotics (Extensively-Drug Resistant, XDR) represents a significant global issue. WHO approximated 558.000 MDR-TB cases, 8.5% which were seen as a XDR-TB [25]. The introduction of MDR and Benzocaine hydrochloride XDR mycobacterial strains provides resulted in the necessity to define brand-new healing choices that, in association with standard chemotherapies, may enhance their effectiveness and represent an additional tool for the control of drug resistance [26]. In this context, the use of adjunctive host-directed therapies (HDTs), which aim to simultaneously limit inflammation and pulmonary damage and boost the host innate antimicrobial response, may represent a stylish avenue deserving further research [26, 27]. Aim of this. A functional understanding of the partnership between glucocorticoids and neuronal apoptosis induced with the creation of reactive air species (ROS) can lead to a book strategy for the procedure or prevention of unhappiness. cytotoxicity in N2A cells. This scholarly research shows that low focus of butein can prevent CORT-induced cytotoxicity in N2A cells, and provides primary results supporting a number of Rabbit Polyclonal to ZNF420 the helpful assignments of butein in neuroprotection. during neuronal advancement because extreme ROS has TAK-875 inhibition a neurotoxic function during neuron differentiation. Furukawa et al. (2019) reported that their synthesized carbazole derivative be capable of protect N2A cells from hydrogen peroxide-induced cell loss of life and induce neurite outgrowth through activation of PI3K/Akt signalling in N2A cells. As a result, we speculate that butein might inhibit CORT-induced ROS era and stop inhibition of RA-induced neurite outgrowth via an intracellular signalling adjustment similar compared to that caused by butein protectivity under CORT-induced apoptosis. Our research demonstrated that butein might induce more apoptotic cytotoxicity in N2A cells in higher concentrations than at 0.5 M. This cytotoxicity considerably affected RA-induced differentiation in the N2A cells at 50 M butein specifically. A focus was utilized by us of 0.5 M butein in a variety of assays as the viability of cells treated with 0.5 M butein hadn’t reduced significantly (Fig. 1A). Chen et al. (2012) reported that butein induced TAK-875 inhibition apoptosis in N2A cells within a dose-dependent way through reduced Bcl-2/Bax proportion and elevated cleavage types of caspase-3 and TAK-875 inhibition PARP. This apoptosis was due to ROS creation at higher butein concentrations relating to our outcomes. In relation to various other neuronal cells, based on the reported MTT assay in HT22 cells (Lee and Jeong. 2016), 10 M butein possess showed no cytotoxic results while an increased focus of 20 M somewhat decreased cell viability. However the reported romantic relationship between cell cytotoxicity and butein dose-dependence continues to be naturally different because of the use of several cell lines and experimental strategies, butein may induce apoptosis in N2A cells in higher butein concentrations under consistent experimental configurations relatively. Chronic tension affects structural changes and neuronal damage in the hippocampus and decreases BDNF in the dentate gyrus (Smith et al., 1995). Chronic administration of several antidepressant medicines significantly improved BDNF mRNA in the hippocampus, and could promote neuronal survival and protect neurons from your damaging effects of stress (Nibuya et al., 1995). BDNF manifestation in brain is known to increase in subjects treated with antidepressants compared with antidepressant-untreated subjects (Chen et al., 2001), and BDNF levels were significantly reduced patients of major major depression (Karege et al., 2002). CREB and BDNF play an important part in neurogenesis and synaptic plasticity in vital areas such as the hippocampus and the cortex for learning, memory space, and cognition (Hashimoto et al., 2004). Improved BDNF manifestation by CREB phosphorylation results in improved secretion of BDNF, which functions via TrkB receptors and activates the MAPK signalling pathway. MAPK signalling phosphorylates CREB and regulates cellular survival by increasing the expression of the anti-apoptotic protein Bcl-2. Cho et al. (2013) reported that the effects of butein on CREB phosphorylation and BDNF manifestation in the hippocampus of scopolamine-induced amnesic mice was identified, and western blotting analysis showed no effect of CREB phosphorylation and slightly improved the BDNF manifestation. Although studies showcased the potent neuroprotective effects of butein, no apparent correlation between the neuroprotective effects and enhancing effects TAK-875 inhibition of butein were found. While TAK-875 inhibition the pharmacokinetic properties of butein are unclear, the bioavailability of flavonoids is definitely low, generally due to limited absorption. In humans, maximum plasma concentrations of polyphenols in the range of 0.1C10 mol/L have been found to be acquired after oral usage, thereafter, the flavonoid in the blood is metabolized extensively and excreted rapidly (Kroon. The affective state is the combination of emotion and feeling, with feeling reflecting a operating average of sequential emotional events with an underlying internal affective condition jointly. distinct, emotional explanations and more and more are thought to possess distinctive neurobiological underpinnings. Do odors possess a similar impact on both moods and emotions? With this manuscript, we briefly review the psychology and biology of feeling and feelings in humans and then focus on links between olfaction, feeling and feelings in the neural circuit and behavioral levels based on data derived from both humans and animal models. The focus is not on whether odors can directly evoke emotions or have hedonic valencethey do and this may be a Rabbit polyclonal to AHCYL2 fundamental adaptation of the olfactory system (Herz, 2000; Yeshurun and Sobel, 2010). Rather the focus is on how odors may modulate ongoing emotional or feeling claims. Finally, we explore how the odor-mood/feelings interplay is definitely, or could be, used in commercial and medical applications. Meanings and Caveats We begin our brief overview of the vast and active field of study on feelings and moods with fundamental meanings necessary Enzastaurin cost to help align neuroscientists and psychologists. For Enzastaurin cost more in-depth meanings and discussions of these terms see evaluations by LeDoux (1996), Russell (2003) and Barrett et al. (2007). The term [or core impact (Russell, 2003)] generally refers to the immediate response to the anticipation or event of rewarding or punishing stimuli or events. Emotions thus tend to become short-lived and event- or stimulus-driven and have a valence (i.e., good or bad). Classic simple (organic) feelings include pleasure, sadness, anger, dread, shock, and disgust. In human beings, feelings are subjective assessments of root physiological and behavioral replies to risk or praise (LeDoux, 2014). Hence, viewing a snake can cause a genuine variety of physiological and behavioral replies to cope with the risk, e.g., activation from the sympathetic anxious program to mobilize energy for get away resulting in speedy center respiration and price, activation of skeletal muscles program to go the body from the risk, or in some cases just the oppositebehavioral freezing. Such physiological and behavioral reactions can be observed in both humans and animal models. However, in humans, overlying these physiological and behavioral reactions is the conscious subjective interpretation of what ones body is doing; resulting, in this case, in the feelings of fear (Russell, 2003; LeDoux and Pine, 2016). A different set of physiological and conscious responses may be evoked upon the sight of ones favorite dessert or true love. An emotional experience (compared to a nonemotional experience) entails the coherent organization of all these components (Russell, 2003; Delplanque et al., 2017). Given the inability to divine conscious, subjective, experiences in nonhuman animal models, there is debate over whether non-human animals display classic emotions in the full Enzastaurin cost sense just described. This review article will not settle that debate. Nonetheless, our understanding of the neurobiology of circuits underlying the physiology of such behaviors is well informed by nonhuman animal research. As noted by Barrett et al. (2007): animal models yield Enzastaurin cost necessary and important insights that must be incorporated into any model of emotion, but they have not (and Enzastaurin cost probably cannot) give a sufficient account of the events people call fear, anger, or sadness (page 298). Thus, our discussion of the neurobiology of olfaction, emotions and moods below relies on both human and non-human animal data. Furthermore, there is some debate over whether there are specific, discrete emotions, e.g., fear, anger, love, or whether emotions fall along continuous dimensions (Panskepp, 1998; Mendl et al., 2010; Hamann, 2012; Lindquist et al., 2013). A variety of models have been developed to describe those dimensions (Russell, 2003; Coppin and Sander, 2016), though most include a valence dimension (i.e., pleasant vs. unpleasant or reward vs. punishment) and an arousal dimension (i.e., high vs. low or intense vs. mild), with the third dimension of potency sometimes also included. Figure 1 shows an example of a two-dimensional emotion plot. In this plot, fear represents relatively high arousal, negative valence emotion, as opposed to excited, which is relatively high arousal but has a positive valence also. For the arousal sizing, both thrilled and relaxing possess an optimistic valence but differ along the intensity similarly.
urinary tract infection nursing care plan Urinary Tract Infection is due to an infection in the lower urinary tract, involving the bladder, and sometimes the urethra and the ureter. It is an ascending infection caused commonly by E. coli, Enterobacter, Pseudomonas, and Serratia. Clients who are at risk for having urinary tract infection are those with long-standing indwelling catheters, pregnant women, females who are sexually active, or those who usually wear synthetic underwear or pantyhose. A client with a urinary tract infection may manifest frequency, urgency, and burning sensation during urination. Some may have hematuria and may experience difficulty voiding. It is usually accompanied by fever and chills and some low back pains. In assessing a client with Urinary Tract Infection (UTI), you may expect the following signs and symptoms: 1. Painful urination 2. Urinary Frequency 3. Fever 4. Flank pain Urinary Tract Infection Nursing Care Plan Nursing Diagnosis Altered Urinary Elimination Possible Etiologies: (Related to) Disease Process ( UTI) Defining characteristics: (Evidenced by) Subjective Data: “I am bothered that I urinate frequently, which would sometimes have discomforts…” verbatim of the client. Objective Data: • Positive urine culture • Feelings of urgency on urination • Some burning sensations when voiding as claimed • Incomplete emptying of the bladder Short term goal: The client will be able to verbalize understanding of the condition and participate in doing measures to compensate for alteration in elimination. Long term goal: The client will be able to achieve a normal elimination pattern as evidenced by decreased in episodes of urgency and frequency in voiding and reduction in discomforts in voiding. Nursing Interventions 1. Obtain a history of illness and factors relating to the condition. 2. Assess the client’s previous elimination pattern and note for the presence of frequency, urgency, burning sensation, size and force of the urinary stream. 3. Determine the presence of pain noting for location, duration, intensity. 4. Monitor fluid intake and output and encourage fluid intake up to 3,000cc- 4,000cc or as tolerated and indicated. 5. Emphasize not to hold urine and assist in toileting routines. 6. Emphasize good hygiene and other measures such as wiping front to back after urination, and voiding immediately after intercourse. 7. Encourage the client to verbalize feelings and concerns relating to the condition. 8. Monitor urine cultures and blood studies. 1. It could aid in providing baseline data regarding the client’s status. 2. It could aid in determining the degree of alteration; helps monitor the progress of the condition. 3. Presence of pain indicates the severity of the condition; it may also indicate the presence of stones increasing the pressure in urination. 4. It could help maintain renal function, prevent from formation of urinary stones, and prevent further infection. 5. Sometimes client holds their urine to avoid experiencing discomforts or just timid to do so; emphasizing not to hold their urine prevents the stasis of bacteria in the urinary system and prevents the formation of renal stone. 6. Hygiene helps in preventing reinfection, especially in women. 7. Open communication could correct factors that contribute to the condition; it could also aid in encouraging the client’s compliance with treatment procedures. 8. It is to monitor the presence of bacteria to determine the presence of infection. Client should verbalize understanding of urinary tract infection, factors contributing to it, and how to manage the illness; participate in measures preventing reinfection. Client should have achieved a normal elimination pattern as evidenced by decreased episodes of urgency, frequency, burning sensation, and other discomforts in voiding. Please enter your comment! Please enter your name here
Best answer: How do I convert a pattern to a vector in Illustrator? How do you save a pattern as a swatch in Illustrator? Select your pattern swatch, go to the arrow in the right of the Panel and select Swatches Library Menu > Save Swatches. Name your pattern and make sure it’s saved under the “Swatches Folder” in an . ai format. How do you fill a vector in Illustrator? Illustrator also allows you to remove the fill from the object. 1. Click the “Selection Tool” or “Direct Selection Tool” icons, which are the black or white arrows in the Tools panel, and then click the object you want to fill. … How do you make a seamless pattern in Illustrator? How to Make a Repeating Seamless Pattern in Illustrator 1. In your document, select the rectangle tool and create a 600px x 600px box. … 2. Start arranging your elements. … 3. Click the square in the background and click Edit>Copy, then Edit>Paste in Front, and bring that copy to the front (Object>Arrange>Bring to Front). How do I fill an object with a pattern in Illustrator? THIS IS FUN:  Question: Can gimp open AI files? How do you rasterize a pattern? Click the Effects menu in the top horizontal toolbar, after you have selected your objects. Select “Rasterize” from the Effects options. Select the Rasterization options, exactly as you would if you were permanently rasterizing an object. How do you break apart a vector image in Illustrator? Go to Modify>Symbol>Break Apart or in Flash Modify Break Apart. I am afraid you are very confused and need a drink. In Illustrator if you have a symbol placed on the canvas then you can break it from the symbol link and make it regular art using the symbils panel to break the link.
Solving email problems Protection against phishing What is phishing? Phishing is an online fraud technique that aims to obtain a person’s personal information in order to steal their identity. How can I recognize a phishing attempt? You’ll receive an email asking you for information like your username, password or date of birth. The email often contains an attachment or a hyperlink to a fraudulent site. If the message appears to have been sent by a well-known company (such as your bank), be extra wary: most institutions don’t request personal information from customers by email. How can I protect myself from phishing? There is no automatic surefire way to protect yourself from phishing. The solution is to take precautions:  - Don’t answer emails that request urgent or immediate action and warn of imminent consequences, don’t click the suggested links, don’t open the attachments and don’t share any personal information.  - Check the quality of the writing: Phishing emails often contain spelling mistakes.  - Before making an online transaction, always check the reliability of the company’s website. What security measures are being taken by Videotron to protect its customers?  - Emails sent to your Videotron address ([email protected]) go through our spam filter, so most phishing emails are intercepted before they end up in your inbox.  - If you receive a message that seems fraudulent, flag it by clicking the “Spam” button in the Webmail section.  - We take your reports very seriously and will take action to block hackers and fraudsters.  - Our Maximum Security Services offer you additional protection against fraud attempts with a system that detects suspicious links sent by email or instant messaging. Has this article answered your questions?
What Are Advanced High-Strength Steels? In many of the world’s languages, there’s a common expression “as strong as iron.” Metals and iron are the first of them and have for centuries been synonymous with hardness and inflexibility. This stable image is quite understandable to us, although it would have been more accurate to say “as hard as steel”— a tough, wear-resistant alloy that copes well with the processing of other metals and abrasive rocks. How Is Advanced High-Strength Steel Made? High strength steel is steel with a tensile strength of at least 1800÷2000 MPa. Trustworthy companies, like Metinvest holding, include the following alloying elements in high-strength steels: chromium, nickel, and molybdenum. These elements significantly increase resistance to brittle fracture. In addition, high-strength steels contain silicon. Silicon considerably delays martensite decomposition and carbide coagulation at low tempering. Silicon alloying makes it possible to obtain high-strength steels by applying higher temperatures and reducing internal stresses. Silicon-containing steels have slightly higher strength at the same temperature, which reduces the carbon content of the steel; that, in turn, improves weldability and increases resistance to brittle fracture. Silicon delays the breakdown of martensite under the influence of plastic deformation. As a result, silicon significantly enhances the mechanical properties of high-strength steels subjected to hardening and low tempering. Types of High Strength Steels Essential characteristics of high-strength steels, in addition to high tensile strength and yield strength, are crack resistance and retention at an acceptable ductility level. Weldability is also an important consideration. The high-strength steels are of the following types: • Medium-alloyed low-temperature steels • Martensitic steels • TRIP steels Advantages of High Strength Steels When it comes to loading, shredding, compacting, and further processing scrap metal, wear and tear on the production equipment is an inevitable part of every stage of the process. Grapples, hydraulic shears, shredders—all these mechanisms, combining directly with the recycled material, need quality wear parts in the first place. Wear-resistant steels have a well-balanced chemical composition and, as a result, are resistant to cracks and dents. Advanced high-strength steels protect equipment and increase its productivity as a whole. Modern machine tools allow the processing of steel at all stages of the production process: cutting, bending, machining, and welding. In this case, it’s possible to repair and restore ready assemblies and equipment and order new ones according to drawings. Proper selection of steel grades, machining with well-established technological production schemes will ensure that the steel will withstand the heaviest loads and last even in the most adverse conditions. Prospects for High-Strength Steels Specialists link the possibilities of further increasing the strength of steels primarily with nanocrystalline materials and materials with highly dispersed nanoscale separations. It has long been known that a reduction in grain size leads to an increase in the strength of metals, but there’s also a decrease in flexibility. However, recent studies have shown that forming a nanocrystalline structure in some metals can significantly increase strength while maintaining high ductility.
Military Extraterritorial Jurisdiction Act (MEJA)  Introduction to the Military Extraterritorial Jurisdiction Act (MEJA)   Jurisdiction means the power of a court to try and determine a case and to render a valid judgment. Courts-martial are courts of special and limited jurisdiction. For example, courts-martial jurisdiction applies worldwide but is limited to a specific class of people— armed forces members. In general, three prerequisites must be met for courts-martial jurisdiction to vest. They are:  1. jurisdiction over the offense,  2. personal jurisdiction over the accused, and  3. a properly convened and composed court-martial.  YouTube video Whether a court-martial is empowered to hear a case—whether it has jurisdiction—frequently turns on issues such as the status of the accused at the time of the offense or the status of the accused at the time of trial. These issues of courts-martial jurisdiction relate to either subject matter jurisdiction (jurisdiction over the offense) or personal jurisdiction (personal jurisdiction over the accused). Subject matter jurisdiction focuses on the nature of the offense and the status of the accused at the time of the offense. For example, subject matter jurisdiction is satisfied if the offense is chargeable under the Uniform Code of Military Justice (UCMJ) and the accused is a servicemember when the offense is committed. Personal jurisdiction, however, focuses on the time of trial: can the government court-martial him? The answer is yes, so long as the accused has proper status, i.e., the accused is a service member at the time of trial.  Sources of Jurisdiction under the Uniform Code of Military Justice (UCMJ) 1. The Constitution: Article I, section 8, clause 14  2. UCMJ, articles 2, 3 and 36  3. MCM, 2005 ed., RCM 201 – 204  4. Customary international law and treaties  Five Elements of Court-Martial Jurisdiction, RCM 201(b):  1. Proper jurisdiction over the offense (subject matter jurisdiction). 2. Proper jurisdiction over the person (personal jurisdiction). 3. Properly composed court (military judge and members must have proper qualifications.) Absent evidence of coercion or ineffective assistance of counsel, the accused’s request to be tried by a military judge alone can be inferred from the trial record (applying “substantial compliance” doctrine to Article 16. United States v. Turner, 47 MJ 348 (1997). Article 25 (request for enlisted members to serve on panel) is also satisfied by substantial compliance. 4. Proper convening authority. A properly constituted court-martial may try any person subject to the UCMJ, even if the accused is not under the command of the convening authority. 5. Properly referred charges. Subject matter jurisdiction is established by showing military status at the time of the offense.  Administrative Double Jeopardy Policies.  Generally, an Armed Forces member will not be tried by court-martial or punished under Article 15, UCMJ, for the same Act a civilian court has tried the Soldier. This policy is based on comity between the federal government and state or foreign governments. See AR 27-10, para. 4-2; JAGMAN, para. 0124.  Jurisdiction in Capital Cases.  Loving v. United States, 116 S.Ct. 1737 (1996). Justice Stevens (concurring) raised whether a “service connection” requirement applies to capital cases. See also United States v. Simoy, 46 MJ 601 (AF Ct. Crim. App. 1996) (a capital murder case in which the court made a specific finding that the felony murder was “service-connected”).  United States v. Gray, 51 MJ 1 (1999). The CAAF gives credence to Justice Stevens’ concurring opinion in Loving. The CAAF makes a specific finding that there are sufficient facts present in Gray, a capital case, to establish a service connection to warrant a trial by court-martial, but does not answer the question of whether a “service connection” requirement applies to capital cases.  Subject Matter Jurisdiction Over Reservists/National Guard.  • The offense must be committed while the reservist has military status. • Jurisdiction attaches at 0001 hours of the effective date of the orders to active duty. • Jurisdiction may exist outside the parameters of the orders. • Offenses committed as part of the accused’s “official duties” may be subject to court-martial jurisdiction, even if the accused is not active.  • United States v. Oliver, 57 MJ 170 (2002). The CAAF found that the medical records submitted on appeal established that the accused had been retained on active duty beyond the expiration of his orders, thus satisfying subject-matter jurisdiction over the offense.  YouTube video A person becomes subject to court-martial jurisdiction upon enlistment in or induction into the Armed Forces, acceptance of a commission, or entry onto active duty pursuant to order. Court-martial jurisdiction terminates upon a valid discharge.  General Provisions: UCMJ, Art. 2, provides jurisdiction over categories of persons with military status:  1. Enlistees; 2. Inductees; 3. Academy Cadets/Midshipmen;  4. Retirees; 5. Persons in custody;   6. P.O.W.’s;  7. Reservists Jurisdiction over retirees is constitutional. Pearson v. Bloss, 28 MJ 376 (CMA 1989); United States v. Hooper, 26 CMR 417 (CMA 1958); Sands v. Colby, 35 MJ 620 (ACMR 1992).  United States v. Huey, 57 MJ 504 (N-M. Ct. Crim. App. 2002). The accused had served 20 years on active duty and was placed on the Retired List on 1 January 1989. In 1996 he worked as a Naval civilian employee in Okinawa. He confessed to sexual intercourse several times a week over nine months with his 16-year old adopted daughter. When the raping stopped, the accused was 58 years old, and his daughter was pregnant with his child. At trial, the accused moved to dismiss for lack of personal jurisdiction based upon a violation of constitutional due process under the Fifth Amendment. The accused cited Toth v. Quarles, 350 US 11 (1955) and argued that he had “obtained civilian status” and was deprived of due process rights available only in a civilian courtroom. The service court disagreed, stating that there “is no doubt that a court-martial has the power to try a person receiving retired pay.” United States v. Stevenson, 65 MJ 639 (N.M.Ct.Crim.App. 2006). Accused was a sailor on the Temporary Disability Retirement List who waived his military disability pay in favor of Veteran’s Affairs disability compensation. Held: Court-martial had personal jurisdiction because the accused was “entitled to pay,” even if he was not receiving pay.  The Article 2(d), UCMJ, involuntary recall process required for members of a reserve component, is not required to bring retirees and members of the Fleet Reserve or Fleet Marine Corps Reserve on to active duty to have jurisdiction over them. United States v. Morris, 54 MJ 898 (N-M. Ct. Crim. App. 2001) petition for review denied, 2001 CAAF LEXIS 597 (22 May 2001).  Involuntary Recall Retired Reservist. Morgan v. Mahoney, 50 MJ 633 (A.F.Ct.Crim.App. 1999). Air Force retired reserve officer was involuntarily recalled to active duty under Art. 2(d)(1). Court held that the accused was subject to court-martial jurisdiction because 1) he was a “person lawfully called or ordered into…duty” under Art. 2(a)(1); 2) he could be ordered involuntarily to AD under Art. 2(d)(1) & (2) for offenses committed while the accused was on AD or IDT (within the statute of limitations); 3) he was amenable to the UCMJ under Art. 3(d) despite the termination of AD/IDT; 4) the AF Reserve is a “reserve component of the armed forces”; 5) he was in a “retired status” under 10 USC 10141(b); and 6) at the time of his recall, he was a member of the Retired Reserve. [Note: Retired Reservists who are receiving hospitalization from an armed force are subject to court-martial jurisdiction without being recalled to active duty].  Persons in custody Jurisdiction terminates once an accused’s discharge is ordered executed (or enlistment expires) and he or she is released from confinement. The remaining suspended punishments are automatically remitted. United States v. Gurganious, 36 MJ 1041 (NMCMR 1993).  Fisher v. Commander, Army Regional Confinement Facility, 56 MJ 691 (N-M. Ct. Crim. App. 2001). An accused that still has military confinement to serve pursuant to a court-martial sentence is still a military prisoner subject to military jurisdiction under the concept of “continuing jurisdiction,” notwithstanding the execution of his punitive discharge and receipt of the DD Form 214. This is true even where the prisoner is serving time in a state civilian prison. The discharge merely terminated his active duty status but did not terminate his status as a military prisoner. YouTube video In a time of declared war or a contingency operation, persons serving with or accompanying an armed force in the field. (covered in more detail in Part VI of this outline)  “Reserve Component” includes USAR and Army National Guard of the United States (ARGUS) soldiers in Title 10, US Code, duty status. (See sections II.E. and IV. of this outline).   Inception of Court-Martial Jurisdiction.  A Contract Which Changes “Status.” UCMJ, Art. 2(b). Art. 2(b) The voluntary enlistment of any person who can understand the significance of enlisting in the armed forces shall be valid for purposes of jurisdiction under subsection (a) of this section and a change of status from civilian to a member of the armed forces shall be effective upon the taking of the oath of enlistment.  Involuntary enlistment:  United States v. Catlow, 23 C.M.A. 142, 48 C.M.R. 758 (1974) (coercion); United States v. Lightfoot, 4 M.J. 262 (C.M.A. 1978); and United States v. Ghiglieri, 25 M.J. 687 (A.C.M.R. 1987) (proposed enlistment as alternative to civil prosecution -no coercion).  Constructive Enlistment:  The codification of In Re Grimley, 137 US 147 (1890). UCMJ, Art. 2(c) (as amended in 1979): Art. 2(c) Notwithstanding any other provision of law, a person serving with an armed force who—  1.  Submitted voluntarily to military authority;  2.  Met the mental competence and minimum age qualifications of sections 504 and 505 of this title at the time of voluntary submission to military authority;  3.  Received military pay or allowances; and (4) Performed military duties;  4.  Is subject to this chapter until such person’s active service has been terminated by law or regulations promulgated by the Secretary concerned.”  Termination of Jurisdiction Over the Person.  General Rule: Discharge Terminates Jurisdiction. ETS/EAS by itself does not terminate jurisdiction.   RCM 202(a) discussion: “Completion of an enlistment or term of service does not by itself terminate court-martial jurisdiction . . . court-martial jurisdiction normally continues past the time of scheduled separation until a discharge certificate or its equivalent is delivered or until the Government fails to act within a reasonable time after the person objects to continued retention.”  RCM 202(c)(1): “Court-martial jurisdiction attaches over a person when action with a view to a trial of that person is taken. Actions by which court-martial jurisdiction attaches include: apprehension; imposition of restraint, such as restriction, arrest, or confinement; and preferral of charges.” See United States v. Self, 13 MJ 132 (CMA 1982); United States v. Benford, 27 MJ 518 (NMCMR 1988).  Ucmjarticle120380 Gonzalez &Amp; Waddington - Attorneys At Law Appellate Leave.  United States v. Ray, 24 MJ 657 (AFCMR 1987) (jurisdiction upheld where accused, on appellate leave, was not provided discharge due to governmental delay in executing punitive discharge).  When is discharge effective?  On delivery.  United States v. Melanson, 53 MJ 1 (2000). Jurisdiction existed because, under AR 635-200, a discharge takes effect at 2400 hours on the date of notice of discharge to the Soldier. See also United States v. Williams, 53 MJ 316 (2000). A valid legal hold had been placed on the accused before the expiration of the date that constituted the effective date of the discharge. United States v. Scott, 11 CMA 646, 29 CMR 462 (1960). A discharge takes effect at 2400 hours on the date of discharge, even if the discharge is delivered earlier in the day (unless it is clear that it was intended to be effective at the earlier time).  Valid Discharge Certificate:  Discharge Authority’s Intent. Early delivery of a discharge certificate for administrative convenience (e.g., the command does not want to keep personnel office open until 2400) does not terminate jurisdiction when a certificate is clear on its face that the commander did not intend the discharge to take effect until later. United States v. Batchelder, 41 MJ 337 (1994). See United States v. Guest, 46 MJ 778 (Army Ct. Crim. App. 1997).  Final accounting of pay.  A final accounting of pay is later than the final appointment at the local finance office. Jurisdiction may still exist several days after a service member has undergone a clearing process and received their DD214. The local finance office is only the first of many steps required to accomplish a final accounting of pay. Undergo a clearing process. United States v. King, 27 MJ 327 (CMA 1989) (sailor refused to complete re-enlistment ceremony after receiving a discharge certificate). Three elements per King to effectuate an early discharge:  1. Delivery of a valid discharge certificate;  2. A final accounting of pay; and  3. Undergoing a “clearing” process as required under appropriate service regulations to separate the member from military service.  Post-arraignment Discharge. A valid discharge of a soldier before trial operates as a formal waiver and abandonment of court-martial in personam jurisdiction, whether or not such jurisdiction had attached before discharge. Smith v. Vanderbush, 47 M.J. 56 (1997). Post-conviction Discharge  a) Effect on Appellate Review and Power of Convening Authority  Execution of Punitive Discharge.  United States v. Keels, 48 MJ 431 (1998). Promulgation of a supplemental court-martial convening order that ordered executed a punitive discharge does not terminate court-martial jurisdiction. Even when there is a punitive discharge, jurisdiction does not terminate until delivery of the discharge certificate and final pay accounting. There is not instantaneous termination of status upon completion of appellate review.  In Personam Jurisdiction in a Foreign Country. United States v. Murphy, 50 MJ 4 (1998). The accused was convicted of premeditated murder and sentenced to death for murders he committed while stationed in Germany. The accused challenged the jurisdiction of the court-martial. He argued that the military investigators misled the German government to believe that the United States had primary jurisdiction under the NATO SOFA. Based on this information, the German government waived its jurisdiction. Had the German government asserted jurisdiction, the accused could not have been sentenced to death because the Constitution of Germany prohibits the death penalty. The CAAF held that the accused lacked standing to object to which sovereign prosecuted the case. The important jurisdictional question to answer is: Was the accused in a military status at the time of the offense and trial? The court found that the accused was. The case was set aside and remanded on other grounds.  Exceptions to General Rule that Discharge Terminates Jurisdiction Exception: UCMJ, art. 3(a).  1. a person is subject to the UCMJ at the time of the offense;  2. the person is discharged without trial; and  3. the person subsequently re-enters the service and is thus subject to the UCMJ at the time of trial.  False Allegations of Sexual Assault in the Military & Why Are They So Common? YouTube video United States v. Erickson, 63 MJ 504 (AF Ct. Crim. App. 2006). Appellant was convicted of violating a lawful order, rape and sodomy of a female under 12, and indecent acts and liberties with a female under 16. The crimes were committed while he was on active duty in the Army. He was discharged and subsequently enlisted in the Air Force. He was sentenced to a DD and confinement for life with the possibility of parole. Where appellant was on active duty in the Army when he committed misconduct, was discharged and subsequently enlisted in the Air Force, and was on active duty at the time of trial. Here, the court-martial had jurisdiction over the appellant under Article 3(a), UCMJ.  Exception: UCMJ, Art. 3(b), a person is obtaining a fraudulent discharge.  Exception: UCMJ, Art. 3(c) – Deserter obtaining a discharge for a subsequent period of service. United States v. Huff, 7 CMA 247, 22 CMR 37 (1956).  Exception: UCMJ, Art. 2(a)(7) – Persons in custody of the armed forces serving a sentence imposed by a court-martial. United States v. Harry, 25 MJ 513 (AFCMR 1987) (punishment cannot include another punitive discharge); United States v. King, 30 MJ 334 (CMA 1990) (prosecuted after BCD executed but still in confinement).  Exception: UCMJ, Art. 3(d) – Separation from Active Components to Reserve Status. Leaving a Title 10 status does not terminate court-martial jurisdiction.  Exception: Intent of the Discharge Authority – When the command places a hold on the accused before 2359 on the date of discharge, even though the discharge certificate had been delivered earlier that day, the discharge does not terminate jurisdiction. Jurisdiction Over the Reserve Component.  Army policy states that Reserve Component soldiers are subject to the UCMJ whenever they are in a Title 10 status: Inactive Duty Training (IDT), Active Duty Training (ADT), Annual Training (AT), or Active Duty (AD). See AR 27-10, para. 21-2.  UCMJ, Art. 3(d). Prevents the termination of court-martial jurisdiction over a member of a Reserve Component who violates the UCMJ while in a Title 10 status by the member’s release from active duty or inactive duty training. Closes jurisdiction gaps recognized by Duncan v. Usher, 23 MJ 29 (CMA 1986).  Procedures and Restrictions: AR 27-10, Chapter 21 establishes procedures for taking punitive action (Art. 15, court-martial) against RC Soldiers.  Procedure: Involuntary Recall to Active Duty. UCMJ, Art. 2(d), authorizes a member of a Reserve Component, who is the subject of proceedings under Articles 15 or 30, UCMJ to be ordered involuntarily to active duty for Article 32 investigations, trial by court-martial, and nonjudicial punishment.   Restrictions on the involuntary recall process.  A member may only be ordered to active duty by an active component general court-martial convening authority (GCMCA). UCMJ, Art. 2(d)(4); AR 27-10, para. 21-3.  Unless the appropriate Service Secretary approved the order to involuntary active duty, the member may not be:  1. sentenced to confinement;  2. forced to serve any punishment involving restriction on liberty except during a period of inactive duty training or active duty; or  3. placed in pretrial confinement. UCMJ, Art. 2(d)(5).  General and Special Courts-Martial. Before arraignment, the reservist must be on active duty. R.C.M. 204(b)(1).  Summary Courts-Martial. Can be initiated and tried within the reserve structure and without active duty involvement. RCM 204(b)(2). But the summary court-martial officer must be placed on active duty. UCMJ, Art. 25; RCM 1301.  Impact on the National Guard.  • 32 USC § 505 – Training in a state status – No federal military jurisdiction.  • 10 USC § 672 – Training in a federal status – Guard member is subject to jurisdiction and the reserve jurisdiction legislation’s major provisions. This includes involuntary recall. But see In United States v. Dimuccio, 61 MJ 588 (AF Ct. Crim. App. 2005) (holding that a Guard member in Title 10 status was not subject to an inspection under MRE 313 ordered by a commander in Title 32 status and suppressing the positive urinalysis resulting from that inspection).  •  The federal status continues until the guard member has completed his federal service (excluding AWOL time), and federal jurisdiction exists notwithstanding state action to terminating jurisdiction. United States v. Wilson, 53 MJ 327 (2000).   1. Pleading Jurisdiction. See RCM 307(c)(3) Discussion at (C)(iv) and (F).  2. Lack of Jurisdiction: Raised by Motion to Dismiss, RCM 907. It may be made at any stage of the proceeding. 3. The burden of Proof. Although RCM 905 states that the Burden of Proof in a motion contesting jurisdiction is a preponderance of the evidence, the government must prove jurisdiction beyond a reasonable doubt if contested at trial.  4. United States v. Bailey, 6 M.J. 965 (N.M.C.M.R. 1979); R.C.M. 905(c)(1)(preponderance); R.C.M. 905(c)(2)(B) (burden of persuasion on government).   Jurisdiction over Civilians  MEJA. Military Extraterritorial Jurisdiction Act of 2000, 18 USC § 3261, Pub. L. No. 106-523.  The MEJA was approved by Congress and signed into law by the President on 22 November 2000. This legislation does not expand military jurisdiction; it extends federal criminal jurisdiction over certain civilians (DOD employees, contractors, dependents, and military dependents) accompanying the military overseas. The implementing regulations went into effect on 3 March 2005. The Act was amended in 2005 to cover civilian employees, contractors, and contractor employees of any Federal agency “to the extent such employment relates to supporting the mission of the Department of Defense overseas.” See 2005 NDAA, Sec. 1088.  The Act applies to felony-level offenses that would apply under federal law if the offense had been committed within the “special maritime and territorial jurisdiction of the United States.”  The Act provides for an initial appearance proceeding, which may be carried out telephonically, conducted by a federal magistrate judge. At this proceeding, the magistrate will determine if there is probable cause to believe a crime was committed and if the person committed it. If pretrial detention is an issue, the magistrate will also conduct a detention hearing as required by federal law. This detention hearing may also be conducted telephonically if the person so requests.  Ucmjarticle120362 Gonzalez &Amp; Waddington - Attorneys At Law Patriot Act. Uniting and Strengthening America by Providing Appropriate Tools Required To Intercept and Obstruct Terrorism (USA PATRIOT ACT) Act of 2001, Pub. L. No. 107–56.  One reason there was a jurisdictional gap before MEJA was that the definition of “special maritime and territorial jurisdiction of the United States,” (SMTJ) was interpreted as excluding US military installations overseas. See United States v. Gatlin, 216 F. 3d 207 (2d Cir. 2000). In 2001, the Patriot Act amended the definition to include military installations overseas. However, the definition excludes anyone already covered by the MEJA. See 18 USC § 7.  Court-martial Jurisdiction under Amended Article 2(a)(10), UCMJ.  The 2007 National Defense Authorization Act amended Article 2(a)(10) as follows:  • “Contingency Operation,” 10 USC Sec. 101(a)(13): The term “contingency operation” means a military operation that- 1. Is designated by the SECDEF as an operation in which members of the armed forces are or may become involved in military actions, operations, or hostilities against an enemy of the United States or an opposing military force; or  Current operations in Afghanistan and Iraq meet the definition of “contingency operation” above.  One civilian tried by court-martial using Article 2(a)(10) jurisdiction. In United States v. Ali, the accused, a Canadian/Iraqi citizen, pled guilty to three specifications involving possessing, hiding, and lying about a knife (the original charge was aggravated assault for stabbing another interpreter in the chest), and was sentenced to five months confinement. His case did not receive an automatic review by the Army Court of Appeals because it was subjurisdictional. Scroll to Top
Freyja's Friday - Sweyn Forkbeard Freyja's Friday - Sweyn Forkbeard England once had a Viking king, But he only reigned for 5 weeks! Sweyn Forkbeard, was the father of King Cnut and in 1013 became the first Viking king of Britain. After a long struggle to get there. He conquered all the thanes of Wessex one by one amassing a huge army as he went, eventually the current King, Aethlred and his two sons abandon the throne and retreated to Normandy and into exile, there upon Sweyn took the throne and became king on christmas day 1013, now that's a present you don’t want to return. Unfortunately he died on 3 February 1014, having ruled England for only five weeks. His embalmed body was returned to Denmark for burial in the church he had built. His son Cnut was pronounced as king. #viking #norse #king  Featured Posts
UniCoronal (Anterior Plagiocephaly) Unicoronal Craniosynostosis Unicoronal craniosynostosis is the second most common type of single suture (non-syndromic) craniosynostosis and occurs when one of the 2 coronal sutures fuses prematurely before birth. The coronal sutures (1 on the right and 1 on the left) separate the forehead bone from the bones at the sides of your child’s head and meet in the middle of the forehead at the anterior (front) fontanelle (soft spot).   Unicoronal craniosynostosis can be associated with other conditions; the Craniofacial Surgeon will examine your child closely. The head shape which occurs because of unicoronal craniosynostosis is known as plagiocephaly which means oblique (slanted) skull in Greek.  Sometimes both coronal sutures are prematurely fused before birth.  When more than one suture is affected, it is called multi suture craniosynostosis. This may happen as part of a syndrome (collection of symptoms) and is known as syndromic craniosynostosis. How common is unicoronal craniosynostosis? It is the second most common type of premature suture fusion affecting approximately 1 in every 10,000 births with a 2:1 female: male ratio which means that it is 2 times more common in girls.  What causes unicoronal craniosynostosis?  The cause of unicoronal craniosynostosis is not yet known. Unicoronal craniosynostosis seems to affect more females than males but we are not yet sure why this is the case.  There may be a genetic basis to the condition as it seems to be passed on from parent to child in a small number of families and in approximately 20-30% of patients, we may be able to identify the gene responsible. When some children are born with unicoronal craniosynostosis a known syndrome (collection of symptoms) can be identified.  Unicoronal craniosynostosis can occur as a part of craniofacial syndromes such Muenke and Craniofrontonasal Dysplasia. More research is needed to identify all the causes of unicoronal craniosynostosis. What are the symptoms of unicoronal craniosynostosis? Some children with unicoronal craniosynostosis may also have a speech and language delay where they tend to start to speak later than other children, this usually improves with help from a speech and language therapist.  Raised intracranial pressure (ICP) seems to develop in approximately 14% of children often when they are between 3 and 5 years of age. Raised intracranial pressure means the pressure inside the skull around the brain increases, which can cause pressure on the brain itself. If your child reports having headaches and are otherwise well with no other signs of an infection such as temperature, sore throat or ears or has a cough it is important to discuss this with a member of the craniofacial team as it may be a sign of raised intracranial pressure.  What does unicoronal craniosynostosis look like? The main sign of unicoronal craniosynostosis is the flatter appearance of the forehead and eye socket on the affected side and the prominence (protrusion) of the forehead on the opposite side. This protrusion on the unaffected side can also cause the eyelid to look as if it is hanging downwards (droopy).   The eyes often look different, with one being larger or ‘more open’ than the other.      When looked at from above your child’s head shape will look flat on the affected side as their forehead slants backwards and it will look bigger (protruding) on the opposite side of the forehead. The affected coronal suture itself may be easily seen as a prominent bony ridge extending from the ‘soft spot’ to the side of their forehead on the affected side. Also, your child’s eye will look more prominent on the affected side because the bones (orbital) surrounding the eye are recessed (sunken) and your child may look amused or surprised. As the eye on the affected side is not in a normal position in the eye socket, it may cause a squint (strabismus) to develop. A squint will cause the eyes to look in different directions such as turning inwards, outwards, upwards or downwards while the other eye looks forwards. Children with unicoronal craniosynostosis must be seen by an Ophthalmologist (Eye Specialist) who will examine your child closely for this. If a squint is not treated it can affect the development a vision. The root of the nose (where the nose is attached to the forehead) may also seem to be tilted (skewed) towards the unaffected side A child with Right sided Unicoronal Craniosynostosis.                                 Does unicoronal craniosynostosis affect the brain? Unicoronal craniosynostosis is associated with brain growth and development problems in approximately 5% of children leading to cognitive (intellectual) and behavioural (developmental and communication) difficulties.  Your child should be assessed by a team of specialists such as a Craniofacial Surgeon, Neurosurgeon and Ophthalmologist (Eye Specialist) to monitor for signs of raised intracranial pressure as this can happen in 14% of children with unicoronal craniosynostosis, especially when they are between 3 and 5 years of age. Does my child need an x-ray? We do not recommend x-rays to diagnose unicoronal craniosynostosis. The Craniofacial Surgeon and Neurosurgeon will make a diagnosis with careful assessment of your child’s skull. However, a CT-Scan with 3D reconstruction (3D CT scan) may be used to: • Verify the diagnosis of unicoronal craniosynostosis • Assess brain growth. • Evaluate if there is evidence of raised intracranial pressure • Assist in surgical planning  Image of a 3D CT scan here What happens in the craniofacial clinic? You and your child will meet the Craniofacial Surgeon, Neurosurgeon, Craniofacial Nurse Specialist and Craniofacial Coordinator. The team will ask you about the pregnancy and your child’s medical /surgical history to date. The team will confirm if your child has unicoronal craniosynostosis and discuss the options available to you. Your child will be referred to an Ophthalmologist (Eye Specialist) and a 3D CT Scan may also be arranged. Once these assessments have taken place you will once again meet with the team at the craniofacial clinic where they will discuss any further questions you may have and the options for surgery to correct your child’s head shape. If your child has other medical conditions, your craniofacial consultant may decide to delay surgery until some of these problems are treated, this is to ensure your child is fit and well for surgery.  Your expectations and conservative management (no surgery) will also be discussed at this visit.  Once a decision for surgery has been made the Craniofacial Nurse Specialist will meet you in clinic to discuss the pre-operative (before surgery) planning including any further assessments or blood tests which must be completed. The Craniofacial Nurse Specialist will also discuss your child’s admission to hospital before surgery including preparation for surgery, expected length of stay in hospital and your child’s expected post-operative (after surgery) recovery course including managing their pain and care of their wound (scar) and the follow up appointments you will need to bring your child to after surgery.  The Craniofacial Coordinator will liaise with all services both inside and outside of Children’s Health Ireland (CHI) at Temple Street to ensure that your child’s entire pre-operative plan is completed before surgery. The Craniofacial Coordinator will also be able to help you connect with other families who have recently gone through similar circumstances and surgery after their child was diagnosed with unicoronal craniosynostosis. The Craniofacial Coordinator will also give you the information about your child’s date of surgery. Will my child need surgery? The decision for surgery is based on severity of the craniosynostosis, the findings of the Craniofacial team with input from other teams where necessary and parental involvement. All families are offered the option of surgical management and treatment for their child’s head shape. For the more severe cases surgery to correct your child’s head shape is almost always necessary. For mild to moderate cases surgical management and treatment is offered. For the cases that are considered mild often conservative (no treatment) management with follow up as per our protocol is indicated.  Early surgery is usually recommended for children with unicoronal craniosynostosis because of the effects on the bones and other features of the face and jaws. If left untreated, unicoronal craniosynostosis may result in a severe abnormality (defect) of the growth of the forehead, orbits and nose which may become more noticeable as your child grows. Whilst the severity of the condition may not worsen, it does become more obvious as the child loses there ‘babyface’ and ultimately grows into a teenager, where the asymmetry can often be very obvious. If it is very severe it may also affect your child’s occlusion (the way the jaw and teeth fit together) and possibly how they chew their food, along with their speech and growth of their lower face.  The team is always available should families wish to discuss surgery in the future if their child’s head shape becomes a problem for their child particularly from a psychological and or social perspective. It is important that families understand the only treatment to change your child’s head shape to a normal head shape is with surgery.  What will my child’s surgery involve? There is a close relationship between the growth of the skull and facial growth. Children with abnormal skull growth will have abnormal facial growth. Since the moulding and shape of the skull is directly dependent on the growth of the brain the best appearance changing results are obtained with early surgery. Surgery for craniosynostosis starts with the Craniofacial Surgeon making a cut in the skin across the top of your child’s head from ear to ear. This cut is shaped like a zigzag which makes it easier for your child’s hair to grow over it after surgery. The Craniofacial Surgeon will decide which parts of the skull bone need to be removed and reshaped (reconstructed) to give your child a normal head shape. Once all the bones are in place the Craniofacial Surgeon will stitch your child’s skin together  Surgical correction of unicoronal craniosynostosis involves a procedure called a Fronto-Orbital Advancement / Anterior Cranial Vault Remodelling (FOA/ACVR) generally; this surgery is performed when your child is between 15 and 18 months of age and is healthy and strong. It is during this time frame that the bones of the skull are thick enough and yet still malleable (pliable) to reconstruct. This surgery involves the removal of the frontal bones (forehead) and of the supra orbital rims ( eyebrow bones) and the fused unicoronal suture. Once the bones are reconstructed the forehead and eyebrow bones are brought forward to allow the brain to grow.  The skull bones are held in place with plates and screws (usually bio-resorbable ones which gradually dissolve between 12 and 18 months after surgery). Usually, there is not enough bone to reconstruct the entire skull and several “soft spots” or defects will remain. These will heal spontaneously over time, up to 24 months after surgery the brain is still growing rapidly, and any soft spots or defects left after the reconstruction will be filled in naturally by new bone being laid down. The surgery takes about 5 hours from start to finish, this includes putting your child to sleep and waking them up. Your child will stay in hospital for approximately 4 to 5 days following surgery.   A child with right sided Unicoronal Craniosynostosis Following surgery, infants and children generally stay in St Gabriel’s ward.  Before going home, the Craniofacial Nurse Specialist will discuss your child’s discharge and home care with you. https://www.craniofacial.ie/wp-content/uploads/2018/08/3.Following-Craniofacial-Surgery.pdf
We are exposed to toxins every day, from the air we breathe to the food we eat, and everything in between. This means the body is working around the clock to rid toxins and prevent their buildup. So, how do toxins leave the body naturally? This is achieved through the work of several organs including the: • Liver filters the blood  • Lungs filter toxins from the air, such as cigarette smoke and pollution • Intestines destroy parasites and other harmful organisms • Kidneys filter excess toxins and waste and release them through urine Since the liver is one of the body’s largest organs, it’s considered the most vital when it comes to detox treatment. The liver is able to break down toxins by changing ammonia to urea, eliminating excess waste from the breakdown of red blood cells, and producing immune cells to eliminate bacteria and harmful toxins. It also processes nutrients and contains thousands of lobules, which produce and release bile from the body.1  The body’s natural detox system is constantly working. However, when there’s an influx of toxins, the organs may not be able to filter as quickly or effectively. Factors such as dietary imbalances and problems due to the growth of cancerous tissue can impair your body’s ability to remove the toxins. When there’s toxin overload, it can weaken your immune system and negatively impact your overall health. Fortunately, there are several types of treatments and methods that promote detoxification and give your organs the filtering support they need. While we can’t avoid toxin exposure, there are ways to clean the blood, prevent toxin buildup, and create stronger immune support and overall well-being. To learn more, we’ve also written in detail answering the question, “what is blood purification?”  IV Chelation Therapy IV chelation therapy is an alternative type of detoxing treatment. It involves the infusion of chemical substances that bind to excess toxic substances like heavy metals and minerals found in the bloodstream and the removal of them from the body. The treatment may involve an IV drip with ethylene diamine tetra-acetic acid (EDTA), a synthetic amino acid that helps to remove common environmental toxins absorbed by the body, such as lead, copper, and iron.  Another route for this type of detox treatment may include an oral supplement of vitamins and minerals or an IV infusion. IV chelation therapy and other alternative detox treatments vary based on your specific needs and symptoms.  Juice Cleanses The central idea behind a juice cleanse is to feed the body adequate amounts of fiber, vitamins, and minerals through drinking the juice from fruits and vegetables. For many, a short-term juice cleanse is meant to work as a reset to their diet followed by incorporating whole, blood cleansing foods, high in antioxidants, into their detox diet more regularly. Reading food labels can help to limit the number of processed food toxins being ingested every day. While a juice cleanse doesn’t remove toxins from the body, it does limit the amount of them being ingested through junk food, sodas, and other processed foods high in sodium and sugar. It provides some relief to the organs and gives them a chance to catch up on filtering out all the other toxins from the environment.2 Activated Charcoal  Activated charcoal is a natural remedy used to remove harmful toxins from the body. Previously, activated charcoal was mainly used by physicians to reduce the adverse reactions of certain poisons. In today’s world, this ingredient has spiked in popularity and has been added to drinks to help detoxify the body. Claims are that charcoal binds to the toxic load in your digestive system and removes them from the blood. It’s important to note with this type of detoxification method that certain medications may lose their effectiveness and too much activated charcoal can cause adverse side effects. Though there are many that tout the benefits of activated charcoal, it’s always best to seek medical advice prior to consuming it.3 Detox Tea Similarly, detox teas have also seen a rise in common use as an alternative detox treatment method. These teas contain specific herbs, such as dandelion, that are known for their diuretic purposes. Since toxins leave the body through urine, these teas can help with the detoxification process in that way, though drinking standard green tea or black tea can also do the trick.  Additionally, proper hydration with water can also keep the organs in good working shape and allow the body’s natural filtration process to continue uninterrupted. Though they can help you feel better, detox teas or a temporary juice cleanse won’t offset the number of toxins the body is susceptible to on a daily basis. Colon Cleanse A colon cleanse is a way to improve digestive tract health. If you regularly experience constipation or irregular bowel movements, it prevents how often and how many toxins leave the body. A colon cleanse can be administered by a doctor, if necessary, or it can be done at home with the right combination of water intake and high-fiber foods. One suggestion is to drink six to eight glasses of lukewarm water per day and eat foods high in water content like watermelons, tomatoes, and lettuce. Another option is to increase the number of fibrous foods in your diet, such as nuts, grains, and other plant foods. This in combination with eliminating foods and beverages high in sugar, salt, and preservatives can help clean out the colon.  It’s a short-term approach to aid toxins in leaving the body without adding more. As with other detoxification methods, colon cleanses aren’t for everyone and should be guided by a doctor to get the best results and err on the side of what’s safest and most effective for your health. Why Detoxification Is Important There’s no question that regular detoxification is important for your overall health. Though your body is designed to naturally remove toxins from the body, there’s still the question of how do toxins leave the body when organs aren’t functioning correctly?  Different detox therapies can help clear the body from the lingering harmful bacteria, germs, heavy metals, and other pollutants that work their way into your system and stay there. Too many toxins can cause symptoms such as: • Feeling sluggish or tired all the time • Bloating and poor digestive health • Constipation or irregular bowel movements • Trouble focusing • Problems sleeping • Mood changes In addition to speaking with your doctor about a detox method that works for you, it’s also important to protect your organs so they can do their job well, especially the liver. Protecting the Liver from Toxins Detox treatment is a helpful way to maintain good health and rid your body of the overload of toxins it’s exposed to. However, it’s also beneficial to practice everyday health habits that help protect the liver and allow it to function at its best.4 A few of these ways include: Eat a balanced diet. Juice cleanses and detox foods and drinks can be temporary ways to help you on your way to optimal health, but overall, eating right every day is what’s best for your body. Adjust your diet to include healthy fibers and fats and avoid processed foods and refined carbohydrates. Also, make sure to drink plenty of water throughout the day, which naturally helps flush toxins from the body.  Be mindful of the products you use. Although it’s impossible to avoid all toxins, using all-natural cleaners, beauty products, and other household items can help limit the number of toxic chemicals and heavy metals you’re exposed to everyday. Toxins enter the body by consuming certain foods and drinks, but also through air pollution and other products that are prevalent in regular life.  Use alcohol in moderation. Alcoholic beverages can damage or destroy liver cells and deter their filtering functionality. Consuming too much alcohol can also cause fat to build up in the liver and lead to symptoms such as inflammation, pain, and fatigue. Since the liver is a key component of the body’s detox system, it’s essential to keep it in good working order. Harmful toxin overload can wreak havoc on your body and can be especially damaging if you need a strong immune system to fight off chronic disease or cancer. Though the body does an exceptional job of natural toxin elimination, a specialized detoxification treatment may be necessary depending on the symptoms you’re currently experiencing and the effects of toxins on your current health. The best path is to speak with your doctor to put together a plan that’s customized for you, one that addresses your specific needs and is administered in a way that’s most effective. The type and frequency of treatment are different based on each situation. Though these may vary, everyone can benefit from fewer toxins in their life. Sometimes the body needs an extra boost to function optimally. The number of alternative treatments available today allows people to focus on a holistic approach toward a healthier life. Additional questions? Call Immunity Therapy Center today. 1. Healthline.com. How to Clean Your Blood. https://www.healthline.com/health/how-to-clean-your-blood 2. UCLA Health – Explore IM. Juicing: Body Cleansing or Nutrient Depleting? https://exploreim.ucla.edu/wellness/juicing-body-cleansing-or-nutrient-depleting/ 3. Healthline.com. Natural Colon Cleanse: 7 Ways to Try at Home. https://www.healthline.com/health/natural-colon-cleanse 4. American Liver Foundation. Healthy Liver – 13 Ways to a Healthy Liver. https://liverfoundation.org/13-ways-to-a-healthy-liver/ August 26, 2021
Cabeza De Vaca Dbq 441 Words2 Pages A journey to remember. Five Spanish ships left the port of Seville in 1527 with 300 people going out to uncharted land called “The New World” and who knew only 4 people would come back. The leader of the entire expedition named Panfilo de Narvaez had dreamed of building settlements on the coast of the Gulf of Mexico. Cabeza de Vaca a military veteran was serving as the treasurer in this expedition. How did he survive? Cabeza de Vaca survived because of his sharp survival skills and because of his success as a healer which lead to his respect for the indians and caused the indians to respect him. Cabeza De Vaca survived because of his wilderness skills and his knowledge of the world moving around him. Cabeza used hollowed out horse legs…show more content… “He once did a surgical operation on a man who got shot with an arrow in the shoulder, with the tip made its way to his heart. Cabeza cut open the man's chest and removed the arrow, then he added stitches using deer bones. Two days later, he came back and he removed the stitches and the man was healed.”(Document C) His medical skills saved the life of the man and it caused him to gain a reputation among the indians. Cabeza De Vaca survived because of his respect for the indians. “Cabeza De Vaca had suffered many disputes. For example, the spaniards who were looking to enslave indians ,and Cabeza was not on good terms with the spaniards because he wanted help the indians, but the spaniards told the indians he wanted to do the complete opposite.”(Document D) This shows the Cabeza De Vaca wanted to stay on good terms with the indians ,but the Spaniards were trying to convince the indians that he wanted to do them harm. Cabeza De Vaca survived with his survival skills, success as a healer, and his respect for the indians. His survival was noble because of the length of the journey and what he had to face. My opinion is that the entire expedition should of been planned out more thoroughly to prevent it from going wrong like it did. Right now mankind is exploring new territory in the amazon rainforest and outer Open Document
What is compartment syndrome? On Behalf of | Jun 28, 2021 | Car Accidents | There are many injuries that someone may suffer in an auto accident. Compartment syndrome is one of them.  Compartment syndrome can be very painful. It involves a buildup of swelling and pressure in the muscles of a person’s arms and legs. This buildup of pressure can affect circulation to those affected body parts leading to a deprivation of oxygen and other essential nutrients your muscles and nerves need to survive. Which body parts are most apt to be affected by compartment syndrome? The body parts most commonly affected by compartment syndrome are a person’s forearms and lower legs; however, their upper arms and legs can be impacted as well. Swelling and pressure along a person’s buttocks, feet, thighs may also occur.  What symptoms do compartment syndrome patients most commonly face? Compartment syndrome patients, in general, experience swelling resulting in blood supply to different regions of their body being cut off. Pain generally results along the portions of the body where the swelling occurs. This can give way to a decreased range of motions and permanent nerve or muscle damage.  Symptoms may vary depending on the type of compartment syndrome at patient has, including:  • Acute compartment syndrome: Generally results from a car accident in which a patient suffers a serious injury like a fractured arm or leg. Permanent damage may result if this injury goes untreated. A surgery (fasciotomy) is generally the best way doctors can relieve the pressure on the muscles. • Chronic compartment syndrome: This is the most common of the two varieties of compartment syndrome and generally doesn’t result in a medical emergency. Patients must instead rest their muscles for the condition to dissipate. This condition is most common among athletes. Compartment syndrome can be quite painful and cause you significant permanent harm if you don’t seek medical attention right away. Your ability to file a personal injury claim in your case may depend on various factors, including the cause of your injuries. Please continue to review our website to learn more about your legal options.  Share This
Cannabigerol (CBG), also known as CBG, is a group of cannabinoids found in the cannabis or marijuana plant. Unlike other cannabinoids, like THC, CBG and its derivatives does not get one high when consumed. Instead, it is reported that CBG blocks the psychoactive effect of THC while still providing the body with medicinal benefits. Part of the CBG group is the cannabinoid cannabigerol monomethyl ether (CBGM).  CBGM was discovered in 1968 by scientists from Kyushu University in Japan. The compound was discovered from a Japanese cannabis strain called Minamioshihara No. 1.  When the scientists were using acid to separate compounds from the cannabis plant, they discovered a pale-yellow colored syrup-like substance which was later called CBGM. They were also able to derive the compound through the process of methylation; this involves the process of adding methyl molecules. CBGM is one of the two cannabinoids in the CBG group that is ether.  At this time, many of the cannabinoids are still undergoing research. Though little is known about this compound (CBGM), the limited knowledge available about it is derived from CBG. There is hope that more research on CBGM will help to fully understand the compound and its effects. Apart from the molecular structure, researchers are yet to know what makes CBGM different from the CBG group. There are possibilities that the compound might affect the body the same way other cannabinoids under CBG does. However, but since CBGM is an ether compound, this could cause it to affect the body in a way different from other cannabinoids. But generally, CBGs are said to interact with the body by impacting the Endocannabinoid System (ECS) of the body. Endocannabinoids are naturally produced cannabinoids by the body that are responsible for the regulation of physiological functions and maintenance of internal balance.  To fully understand the therapeutic effects of CBGM, more research has to be done. One can only speculate that since CBGM is a member of the CBGs, it may provide the same medicinal benefits as other CBGs. If CBGM imitates the therapeutic effects of CGB, then it has potential in helping patients with the following conditions; There are reports that CBG can treat glaucoma by reducing the painful intraocular eye pressure. Although the study has only been done on study rats, scientists are confidents that the compound might have the same effects if used on humans.   Anti-inflammatory effects It was reported that CBG showed excellent anti-inflammatory effects than CBD. The study was done on lab rats with inflammatory Bowel Diseases; ulcerative colitis and Crohn’s disease, and they benefitted from the compounds’ anti-inflammatory therapeutic benefit.  CBG is also reported to have anti-cancer and analgesic benefits. It is also capable of reducing nausea. Another study done on rat models showed that CBG is capable of treating neurodegenerative diseases like Huntington’s diseases.  Cbd Products For Sale Banner COLASANTI, B. K. (1990). A Comparison of the Ocular and Central Effects of Δ9-Tetrahydrocannabinol and Cannabigerol. Journal of Ocular Pharmacology and Therapeutics, 6(4), 259–269. Borrelli, F., Fasolino, I., Romano, B., Capasso, R., Maiello, F., Coppola, D., … Izzo, A. A. (2013). Beneficial effect of the non-psychotropic plant cannabinoid cannabigerol on experimental inflammatory bowel disease. Biochemical Pharmacology, 85(9), 1306–1316 Banerjee SP, Snyder SH, Mechoulam R. Cannabinoids: influence on neurotransmitter uptake in rat brain synaptosomes. J Pharmacol Exp Ther. 1975;194(1):74-81.] than THC.
• The Law Gazette From Presumption to Punishment Presumption means to ascertain some facts on the basis of possibility or the consequences of the situation and circumstances which strengthen the activity happened. Under the law, the Court on the basis of certain facts frames some inferences which are known as presumptions. Those facts are presumed as proved until and unless they are disproved. Presumption can be either affirmative or negative drawn from the circumstances through best possible way of reasoning. Section 326 of the Indian Penal Code, 1860 (IPC) defines the voluntary action of causing grievous hurt which states that any person, who inflict or cause grievous hurt by any means, which is likely to cause death of the person so inflicted shall get punishment for the life imprisonment or imprisonment up to 10 years and fine. Basic essentials that can hold a person liable for the offence under this section are: 1. The person causing grievous hurt must have knowledge about the same. 2. He must be doing the act voluntarily and nobody has forced him to do it. 3. He committed the grievous hurt by the means of any instrument of shooting; stabbing or cutting; any instrument, which if used as weapon can cause death of the other person; fire or any heated substance; any poison or corrosive substance; any explosive substance; by any substance which is difficult to swallow or inhale for the human body or any animal. The other section is 326A of IPC which provides about causing grievous hurt by the use of acid on the other person. It states that “Whoever causes permanent or partial damage or deformity to, or bums or maims or disfigures or disables, any part or parts of the body of a person or causes grievous hurt by throwing acid on or by administering acid to that person, or by using any other means with the intention of causing or with the knowledge that he is likely to cause such injury or hurt, shall be punished with imprisonment of either description for a term which shall not be less than ten years but which may extend to imprisonment for life, and with fine”. This provision says that any person who throws acid on the whole body of another person or any part thereof, which brings deformity in the body, with the intention of causing hurt or injury shall be punished with the imprisonment of 10 years and which may extend to life imprisonment and with fine. And Section 326 B of IPC talks about attempting to cause grievous hurt by throwing acid on the other person for which the punishment is imprisonment for 5 to 7 years and fine. Presumption for crime is given under Sec. 144 of the Evidence Act, 1872 which states that “the court may presume the existence of any fact which it thinks likely to have happened, regard being had to the common course of (a) natural events, (b) human conduct, and (c) public and private business, in their relation to the facts of the particular case”. It means that the Courts have power to presume any fact which it thinks can happen with regard to the natural events, human conduct and public or private business. There are three types of presumptions: 1. May Presume- Here, the Court has the discretion to presume certain facts and circumstances. 2. Shall Presume – Here, the Court has a strong assertion to consider certain facts and circumstances. 3. Conclusive Presumption – Here, the Court has the absolute power to presume and shall not allow any proofs against such presumption. In the last few years, the cases of acid attacks have increased with high rates in India. Earlier acid was being used in households for cleaning purposes but now it is being used to take revenge. Under Section 326A of IPC, Acid includes any substance which is of acidic or corrosive character and has a burning nature that can cause bodily injury leading to the scars, disfigurement, or temporary or permanent disability. Throwing acid is the most vicious crime which is being done by people, generally, just because they are not able to handle the rejection and they consider and treat women as a thing and not a human being. Land, property, and/or business disputes accounted for 20% of acid assaults in India from 2002 to 2010. Land, property, and/or business disputes accounted for 20% of acid assaults in India from 2002 to 2010 [1]. There are many cases regarding this, like Devanand v. The State [2] where a man threw acid on his wife because she refused to live together with him. She suffered permanent disfigurement and loss of one eye. The accused was punished with imprisonment of 7 years under Section 307. This act results in the imperfection or disfigurement in the body and Mental and physical trauma to the victim and her family. The medical effects of the acid attacks are very extensive. This also affects the social, economical and personal life of the victim and her family. There was no law to deal with this problem till 2012. In 2012 the 18th Law Commission of India proposed new sections to be added in IPC, i.e. Section 326A and 326B which provides about causing grievous hurt by throwing acid and attempting to throw acid on a person, respectively and Section 144B in the Indian Evidence Act, 1872 which provides about the presumption in the case of acid attacks. These amendments were done in order to punish the offenders who are indulging into the activities of throwing acid. Man chose acid as an alternative form of offence to take revenge and ruin the lives of women. Presumption in crime means to presume that certain facts can be true according to the circumstances of the case until they are disproved. Under Section 114 of the Evidence Act, 1872, the Court may presume any fact which it deems fit to have happened, while considering the common course of natural events, human conduct and public and private business. Presumption for any fact can be made but that should be within the ambit of common parlance of the natural events, while considering human conduct and public and private business. Section 114 B of the Act, 1872 provides for the presumption in the cases of acid attacks. It says that if any person throws or attempts to throw or administer acid on the other person then the Court shall presume that the act has been done with the intention or with knowledge that it is going to cause hurt or injury as provided under the Section 326A of IPC. These provisions were added in the Evidence Act and IPC under the Criminal Law (Amendment) Act, 2013 after Mukesh and Ors. v. State for NCT of Delhi [3] (Nirbhaya Rape Case). In the case of the State of Maharashtra v. Ankur Narayanlal Panwar [4], (Preethi Rathi’s Case) the girl refused the marriage proposal of the boy and boy out of jealousy took revenge by throwing acid on her face. The accused was held liable and was punished with the death penalty which was later commuted to imprisonment for life for the offence he committed. The Court said that “In the present case, the Trial Court has not applied its mind to the aforesaid factors. The accused was a young boy aged about 23 years. There is no past criminal record. The mitigating circumstances reflected as referred to in the aforesaid decision, were not considered in proper perspective by the Trial Court. Taking into consideration over all circumstances, we do not find that present case can be termed as "rarest of the rare" case and hence the appellant do not deserve death penalty. Hence, the sentence of death penalty is required to be set aside while confirming the conviction”. These crimes are increasing day by day in the country. So, there is still a need for harsher punishment which can only be done by the Parliament. The legislature should frame such stringent laws that the offender thinks multiple times before committing a crime, then only the number of such offences can be reduced. Therefore, there is the requirement of making more strict laws to fill up the loop holes of the Criminal Justice System. 1. https://www.lawctopus.com/academike/the-corroding-faces-of-indian-women-acid-attack/ (Last Visited on 26th October, 2020) 2. Devanand v. The State, [(1987) 31 DLT 184] 3. Mukesh and Ors. v. State for NCT of Delhi (Nirbhaya Rape Case), [(2017) 6 SCC 1] 4. State of Maharashtra v. Ankur Narayanlal Panwar (Preethi Rathi’s case) [2019(4)BomCR(Cri)143] This blog has been authored by Dipti Jain, who is a 4th Year B.Com., LL.B. student at Jamnalal Bajaj School of Legal Studies, Banasthali Vidyapeeth, Rajasthan.
Federalist vs. Anti-Federalist 21 Aug No Comments Federalist vs. Anti-Federalist  People, Power, and Politics Federalist vs. Anti-Federalist Framing the constitution was no easy feat. Some saw this document as only protecting the rich and having them become too powerful. They also saw it as the majority being able to make all the decisions. The “common” man or minority, viewed this as a threat which then lead to two opposing groups, the federalist and the anti-federalist. The federalist consisted of lawyers, bankers and those who thought the articles of confederation was too vague and could not govern our new nation. Although, the federalist were the wealthy they kept in mind to make sure the minority had some say as well. The Federalist are those who favored a strong national government. While the Anti-Federalists did not want a strong national government, and would rather have strong state government. The anti-federalist wanted to prevent the national government from having too much power to prevent tyranny. The anti-federalist consisted of small farmers, shopkeepers, and other laborers. I, personally, would side with both of the parties. I think that there is a need for a strong federal government to make important decisions but I also believe in local government and allowing the state to decide what is best for its people. In federalist No. 51, James Madison stated, “Federal governments separation of powers would prevent one branch from dominating the national government and also the rights of citizens”. (American Government, 2014 12thEdition) This statement shows that the Federalist party had all people in mind when framing the constitution. At the end of it all, the Constitution was ratified so the federalist won but the anti-federalist were able to have a great influence in the document and to maintain a system of checks and balances that keeps branches from having too much power. The bill of rights also greatly reflects the influences the anti-federalist had with making sure the people maintained certain rights and liberties. In conclusion, I think that both federalist and anti-federalist played a huge role in shaping our government that we currently have today. Their efforts made it possible for us to have years and years of effective leadership and government. This is what makes the United States of America so great today, is that we believe in the freedoms of all people. It was not an easy road, by far, and some still have speculations on how this nation should be ran. Although not perfect, the constitution is an amazing document that has helped our nation become one of the greatest in the world. This is a direct result of both parties, the anti-federalist and federalist. Anti-Federalist vs Federalist. (). Retrieved from https://www.diffen.com/difference/Anti-Federalist_vs_Federalist Click following link to download this document SS 236 Federalist vs. Anti-Federalist.docx
What is the Fastest Insect in the World and Where does it Come From? The Australian dragonfly is the Fastest Insect in the World. A dragonfly is a type of insect with large multifaceted eyes, two pairs of transparent wings, and an elongated body. Dragonflies are similar to damselflies, but the wings of most dragonflies are held away from the body when at rest. They are not capable of walking even though they have 6 legs like any other insect. Dragonflies eat mosquitoes, and other small insects like flies, ants, butterflies, and bees. They are found around lakes and ponds because their larvae, also known as nymphs, live in the water. How fast is a dragonfly? It depends on who you ask. Experts estimate anywhere from a remarkable 35 mph to an unbelievable 60 mph. About Karen Hill Leave a Comment
The Nagas Hill Peoples of Northeast India Project Introduction The Naga Database manuscript notes on the Zemi Nagas by Ursula Graham Bower caption: dances caption: King-king-pun medium: notes person: Zingkaolang/ of Magulong person: Graham Bower/ Ursula date: 1939-1946 refnum: Betts papers, ring binder 1 person: Centre for South Asian Studies, Cambridge text: King-king-pun. text: Two lines are drawn on the ground about 20 inches apart and parallel. Bucks stand behind one, right or left shoulder leading, and the other arm folded across the chest, hand on opposite (leading) shoulder; girls stand behind the other, facing more or less forward, hands crossed behind their backs, and armed with a dab of ash moistened with spittle. While the choir sing as for a dance, the individuals on either side try and score a point, "killing" one of the opposite side. The girls try to get a dab of black ash on the chest of one of the bucks, the bucks to catch her and pull her over on to their side. The strip between the two lines is no man's land, into which players on both sides dart in a feint at one of the enemy. In the game we saw the girls were rather chary of joining in, lest they got their breasts touched, but the assembled audience seemed to think them rather faddy, as this wouldn't matter. As the game proceeds the fun waxes fast and furious, with dabs and leaps and darts and retreats (one may not cross behind the enemies' line) and players "killed" on either side fall out and stand in the wings, others taking their places, and it only ends when everyone is out of breath. text: In the game seen at Magulong, the girls were rather chary of joining in, and of their side only one was actually a girl, the others being Zingkaolang, the "maitre de ballet"; a sprightly grandfather; and another elderly man. The "girls" caught out 2 or 3 of the bucks, their only casualty being Zingkaolang himself, slapped and captured.
Skip to content Cholesterol Reviewed: What Can One Be taught From Different’s Mistakes Diuretics are also doubtlessly dangerous for diabetics, because the substances are known to affect insulin and glucose ranges within the body. Given that loop and thiazide diuretics promote potassium loss, amiloride is given to counter that effect and make sure that potassium ranges don’t fall too low. The rationale for that is that by inhibiting aldosterone-delicate sodium reabsorption, much less potassium and hydrogen ion are exchanged for sodium by this transporter and therefore much less potassium and hydrogen are misplaced to the urine. However, there are lots of unwanted side effects associated with diuretics which may be because of lost water weight and elevated urination. Though all diuretics result in elevated urine output and reduced welterweight inside the physique, they are all prescribed for different subpopulations. The term “calcium-sparing diuretic” is sometimes used to establish agents that lead to a relatively low charge of excretion of calcium. A diuretic provides a means of forced diuresis which elevates the speed of urination. The reduced concentration of calcium in the urine can lead to an increased charge of calcium in serum. To maintain osmotic stability, water is retained within the urine. Diuretics are categorized in line with the renal buildings they act on and the modifications they result in in the amount and composition of urine, in addition to electrolyte steadiness. There are a lot of sorts of diuretics, however most act by reducing the amount of fluid that’s reabsorbed by the tubules of the kidneys, whence the fluid passes back into the blood. Thiazide-type diuretics reminiscent of hydrochlorothiazide act on the distal convoluted tubule and inhibit the sodium-chloride symporter resulting in a retention of water in the urine, as water usually follows penetrating solutes. Because the tubule dives into the medulla, or center zone of the kidney, the tubule turns into narrower and varieties a loop (Loop of Henle) that reenters the cortex because the thick ascending limb (TAL) that travels back to near the glomerulus. This causes extra sodium (and water) to cross into the gathering duct and be excreted within the urine. This transporter usually reabsorbs about 25% of the sodium load; subsequently, inhibition of this pump can result in a significant enhance within the distal tubular focus of sodium, decreased hypertonicity of the encircling interstitium, and less water reabsorption in the amassing duct. First, its exercise depends on the tubular focus of sodium, so that when sodium is excessive, extra sodium is reabsorbed and more potassium and hydrogen ion are excreted. The TAL, which is impermeable to water, has a cotransport system that reabsorbs sodium, potassium and chloride at a ratio of 1:1:2. Approximately 25% of the sodium load of the unique filtrate is reabsorbed on the TAL. Most diuretics produce diuresis by inhibiting the reabsorption of sodium at totally different segments of the renal tubular system. Blocks epithelial sodium channels, inhibiting sodium reabsorption within the late distal convoluted tubules and accumulating ducts. Increased thirst Vasopressin V2 receptor antagonist – lowering the number of aquaporin channels in renal amassing ducts, decreasing water reabsorption. A number of the extra common unintended effects of diuretics are dehydration, dizziness, headache, and increased thirst. We are fast to cite: “together we’re stronger”. The quick-time period anti-hypertensive action relies on the truth that thiazides decrease preload, reducing blood pressure. Nearly the entire sodium initially filtered is reabsorbed by the kidney, so that less than 1% of originally filtered sodium stays in the final urine. That’s, the discount in blood pressure will not be resulting from decreased blood volume ensuing from increased urine production, but happens via other mechanisms and at decrease doses than that required to produce diuresis. Chemically, diuretics are a various group of compounds that both stimulate or inhibit various hormones that naturally happen within the body to regulate urine manufacturing by the kidneys. Diuretic medicine improve urine output by the kidney (i.e., promote diuresis). With that mentioned, you should be getting your kidney functioning examined by a doctor or medical professional on a constant basis when taking diuretics. This is completed by altering how the kidney handles sodium. When a diuretic is taken, it reduces the quantity of fluid in the body and likewise rids the body of excess sodium (among other substances). Hyponatremia is a situation during which your body has low ranges of sodium. Ultimately, mental health apps made by BIPOC for BIPOC might help marginalized individuals discover extra of the assist and assets that they need. Below, we’ve tabulated diuretics pharmacology into three columns – detailing the information about drug lessons, negative effects and clinical components you might want to know. What are the three forms of diuretics? By way of medications, there are three sorts of diuretics: Loop, Potassium-Sparing, Thiazide. You, like me, by now are fed up with OTC gimmicks and their unwanted side effects.
Home » Other » What are disposable gloves made from ? What are disposable gloves made from ? Posted by: Yoyokuo 2020-11-26 Comments Off on What are disposable gloves made from ? Medical gloves are disposable gloves used during medical examinations and procedures that help prevent cross-contamination between caregivers and patients. Medical gloves are made of different polymers including latex, nitrile rubber, vinyl and neoprene; they come unpowered or powdered with cornstarch to lubricate the gloves, making them easier to put on the hands. Cornstarch replaced tissue-irritating Lycopodium powder and talc, but even cornstarch can impede healing if it gets into tissues (as during surgery). As such, unpowered gloves are used more often during surgery and other sensitive procedures. Special manufacturing processes are used to compensate for the lack of powder. There are two main types of medical gloves: examination and surgical. Surgical gloves have more precise sizing with a better precision and sensitivity and are made to a higher standard. Examination gloves are available as either sterile or non-sterile, while surgical gloves are generally sterile. What are disposable gloves made from ? Caroline Hampton became the chief nurse of the operating room when Johns Hopkins Hospital opened in 1889. When “(i)n the winter of 1889 or 1890” she developed a skin reaction to mercuric chloride that was used for asepsis, William Halsted, soon-to-be her husband, asked the Goodyear Rubber Company to produce thin rubber gloves for her protection. In 1894 Halsted implemented the use of sterilized medical gloves at Johns Hopkins.The first disposable latex medical gloves were manufactured in 1964 by Ansell. They based the production on the technique for making condoms. These gloves have a range of clinical uses ranging from dealing with human excrement to dental applications.Criminals have also been known to wear these gloves during the commission of their crimes. These gloves are often chosen because their thinness and tight fit allow for dexterity. However because of the thinness of these gloves, fingerprints may actually pass through the material as glove prints, thus transferring the wearer’s prints onto the surface touched or handled. The participants of the Watergate burglaries infamously did so wearing rubber surgical gloves in an effort to hide their fingerprints. What are disposable gloves made from ? Alternative to latex  Due to the increasing rate of latex allergy among health professionals, and in the general population, gloves made of non-latex materials such as vinyl, nitrile rubber, or neoprene have become widely used. Chemical processes may be employed to reduce the amount of antigenic protein in Hevea latex, resulting in alternative natural-rubber-based materials such Vytex Natural Rubber Latex. However, non-latex gloves have not yet replaced latex gloves in surgical procedures, as gloves made of alternative materials generally do not fully match the fine control or greater sensitivity to touch available with latex surgical gloves. (High-grade isoprene gloves are the only exception to this rule, as they have the same chemical structure as natural latex rubber. However, fully artificial polyisoprene—rather than “hypoallergenic” cleaned natural latex rubber —is also the most expensive natural latex substitute available. Other high-grade non-latex gloves, such as nitrile gloves, can cost over twice the price of their latex counterparts, a fact that has often prevented switching to these alternative materials in cost-sensitive environments, such as many hospitals. Nitrile is a synthetic rubber. It has no latex protein content and is more resistant to tearing. Also it is very resistant to many chemicals and is very safe for people who are allergic to latex protein.Nitrile gloves are the most durable type of disposable gloves. Although nitrile gloves are known for their durability, extra care should be taken while handling silver and other highly reactive metals because those substances can react with sulfur, an accelerant in nitrile gloves. Powder-free medical gloves are used in medical clean room environments, where the need for cleanliness is often similar to that in a sensitive medical environment Link to this article:What are disposable gloves made from ?
A day late and a few hundred billion dollars short — Limiting climate change could save the US a ton of money Inaction would cost at least $500 billion per year by the end of the century. Limiting climate change could save the US a ton of money Despite what the world’s least-interesting talking gecko would have you believe, no one likes handing over payments for car insurance. But there’s one thing everyone likes even less: suddenly paying for expensive repairs not covered by your insurance. Similarly, opponents of action on climate change like to complain about the costs of eliminating fossil fuel emissions. Typically, this implies that the alternative—ignoring climate change—is free. It is not. A new study by Jeremy Martinich and Allison Crimmins of the US Environmental Protection Agency provides the most detailed estimate yet of the economic costs of climate change in the United States. They found that taking action to reduce emissions could save us at least $200 billion per year by the end of the century. The effort involves the use of models to simulate the ways different economic sectors are affected by climate change and how those effects ripple from one sector into others. A notable 2016 study used past relationships between temperature and economic activity to project the impact of warming on six economic sectors in the US, drilling down to the county level. This new study tries to expand that to 22 sectors and relies less on mathematical relationships from the past to project to the new climates of the future. The 22 sectors include fisheries, electrical generation, coastal property, roads, harmful algal blooms, air quality, and others. Each was analyzed with the results of five different climate models simulating two scenarios: a high-greenhouse-gas-emissions “business-as-usual” scenario that warms the world around 4.5°C (8°F) by the end of this century and an intermediate emissions scenario that results in about 2.5°C warming (4.5°F). Everything was analyzed for the lower 48 states, while details on Alaska, Hawaii, and Puerto Rico were limited. The model results show that either greenhouse gas emissions scenario would be costly. In the high-emissions scenario, the calculated economic damages added up to about $500 billion per year by 2090. That is due to things like lost labor and premature deaths caused by changes in temperature extremes and air pollution, loss of infrastructure damaged by flooding, and even recreation and tourism losses from disappearing coral reefs, declining snowpack, and harmful algal blooms. The researchers describe these estimates as conservative, given that a limited set of impacts are included. “For example,” they write, “the wildfire analysis captures only suppression costs, and does not estimate health impacts from degraded air or water quality, property damage, or timber loss.” As in previous studies, there are large differences between the damages in different regions of the US. The impact of extremely hot weather, flooding, and West Nile virus, for example, are significantly greater in southern states. The loss of tourism revenue as coral reefs disappear is obviously a bit more relevant in Hawaii than Oklahoma. And coastal property is more threatened in low-lying Florida than in Alaska, where geologic activity is raising land elevations along the coast. But there are enough factors included here that problems in one region tend to be balanced by different impacts in others. If you compare per-person impacts, the regions actually come out pretty close. The calculated per capita value of reducing emissions, in terms of economic damages avoided. Enlarge / The calculated per capita value of reducing emissions, in terms of economic damages avoided. By comparing the estimated damages in the lower and higher emissions scenarios, you can see the economic value of actions that cut emissions. The difference between the two scenarios, in this case, is a cool $220 billion per year by 2090 for the US as a whole. That’s how much economic damage is avoided by steering toward one future rather than the other—even in the low emissions scenario, there are still $280 billion per year in damages you didn’t avoid. (Free lunches are off the table, but the entrees offered come with varying degrees of food poisoning.) Mitigating climate impacts by limiting warming isn’t the only tool available—we can also adapt to changes as they come, lessening the damage they do. The study also tested some specific adaptation scenarios for a few sectors. This includes strengthening and protecting infrastructure from things like sea-level rise; another example looked at here involves the management of railroads during hot weather. At high temperatures, trains have to slow down because the tracks can buckle as they expand. By investing in a network of sensors to monitor rail conditions, rail companies could avoid unnecessary delays. The researchers ran scenarios without any adaptation or mitigation, with just adaptation, and with both. Damages were smallest when both strategies were employed, which shouldn’t be surprising. But for the examples they analyzed, the adaptations made a huge difference. Damage to coastal property, for example, accounts for $120 billion by 2090 in the high-emissions scenario and still costs $92 billion in the lower-emissions scenario. But add in adaptation and that cost drops to $5.7 billion per year. While adaptation and mitigation comes with a price tag, you’d have to know the cost of inaction to figure out the amount that would actually be lost from your future bank account. While economic analyses commonly have to assign monetary value to things like premature deaths, most people will argue there is also some non-monetary value missing from the ledger. Nature Climate Change, 2019. DOI: 10.1038/s41558-019-0444-6 (About DOIs). You must to comment. Channel Ars Technica
Traditions and Superstitions about funeral and dead people. Religious funeral traditions: what should not be done at funerals? The death of a person is an area of the incomprehensible and sacred. All the unknown things are scary, and fears are always associated with prohibitions and taboos. Whatever the origin of these prohibitions, they have long been rooted in folk culture, signs and norms of behavior. A funeral is a time when even a non-superstitious person tries to observe all funeral rites and traditions, follows signs and beliefs, regardless of how he relates to them. Death occupies special place in any religion, and Orthodoxy is no exception. The priest conducting the funeral service can talk about what the Orthodox can and cannot do at the funeral. Let's list the basic rules. You cannot order the last rites for the unbaptized person. According to the Charter of the Church, it is forbidden to perform the rites of burial and church commemoration of people who are unbaptized, baptized but renounced the faith, those who treated the Church and the faith with mockery or enmity, unbaptized and suicides. For suicides, an exception is possible, but only if the person was mentally ill (which is confirmed by medical certificates) and permission from the diocese was obtained for the last rites. You cannot bury an icon in a coffin with the deceased. Some rectors believe that it is wrong to bury the dead person with the icon. Although many Orthodox Christians agree with this, it is important to know that there is no prohibition on the part of the church on this issue, so those who buried their loved ones along with the icons need not worry. Thus, whether to leave the icon from the coffin after the funeral is a private matter for the loved ones of the deceased. If it serves as a spiritual guide, a connection with this person, a guideline for good wishes for the soul of the deceased, it should not only be preserved, but also regularly obtained for prayer and veneration. If the grief over the loss is so strong that every reminder of the departed causes inconsolable grief, one should not torment oneself, but let go of the memory of the deceased with light heart, putting the icon with the deceased or returning it to the temple. You cannot put icons in the coffin if cremation is to come. Although the icon is placed in the coffin, it must be removed from the coffin before cremation. Burning icons (even during cremation) is blasphemy and a sin. What are the Christian folk omens and beliefs about dead people and funeral? Folk culture contains many rules, prohibitions and restrictions on the subject of funerals. Some of them have good medical reasons; others have long been the rules of funeral etiquette. There are also just superstitions that have long merged with the traditions of funerals. Let's list the main things that cannot be done at a funeral: The deceased person should not lie in coffin with open eyes. There are countless explanations for this. Someone is afraid that the open eyes of the deceased are looking out for someone to take with them and soon someone else will die. Others believe that the deceased no longer needs external vision for this world: by closing the eyes of the deceased, we help him gain internal vision suitable for the other world. The mouth of the deceased must also be closed. You should not put the belongings of living people in the coffin. An ancient folk example says: put your thing in a coffin, you will follow the deceased. You shouldn’t hammer the lid of the coffin in the house of the dead person. Popular wisdom believes that if the coffin lid is hammered in the house of the deceased, then another death awaits his family. Only men who are not kin to the deceased should take out the coffin. It is believed that blood is drawn to blood, and the deceased will take his relatives away with him. Carrying out the coffin, you must not touch the doorframe. It is a bad omen: it is believed that the person who struck the coffin with the deceased will soon go to the other world himself. You should not carry the deceased head first. It is believed that at a funeral, the deceased should not be carried head first. He should be carried only with feet forward, otherwise the soul cannot return back. This ancient pagan belief has long come into use and is known, perhaps, to everyone. There is an omen that when the coffin is carried out with the deceased, the alive should not turn back. Since ancient times, there is a belief that, seeing off the deceased, you cannot look back; otherwise death will come for you or into your house. Leaving the cemetery after the funeral, one must not look back. The superstitions say that you should not clean the house while the body of the deceased is in it. A popular proverb says: “sweep out the litter in the presence of the dead - take everyone out of the house”. Pregnant women and children under three years old are not allowed to attend funerals. This prohibition is associated with the idea that a new life is incompatible with death, and the psyche of children is not yet stable enough and the attributes of death can affect them negatively. You can not come to the funeral in light-colored clothes. In many cultures, the color of mourning is black. It is believed that it is not necessary to wear black, but shades of clothing should be dark: light shades supposedly attract the attention of death. According to superstitions about funeral and dead men, you should not walk in front of the coffin. According to the omen, the one who walks in front of the coffin can "leave" to the other world after the deceased. At a funeral, you also cannot overrun the hearse - this promises illness or trouble. Omens state that one should not move towards the funeral procession. If a funeral procession is moving towards you, it is better to stop and wait. The man must take off his headdress. When meeting with a funeral procession, you must not cross the road, since by doing so you cross the road of death, and it may attract it to you. According to the superstitions, at the cemetery, you must not make noise, swear or argue. The soul of the deceased may not like it, and it will be bad for him in the next world. But the point is, rather, not in belief, but in what was said at the beginning of the article: death is scary, and everyone becomes silent. As omens state, you should not kiss the deceased person on the lips. Saying goodbye to the late loved one, you need to kiss the corolla on the forehead, but you should not touch the body with your lips. This sign also has a practical side: in this way you can pick up an infection. According to the superstitions, items should not be taken from the cemetery. You can only take what you have with you, for example, your personal belongings. Everything else, for example, the handkerchiefs and towels used during the funeral, must be left, otherwise you will take death with you. During a funeral and memorial service, one cannot speak badly about the deceased. Each person has something good and something bad. If you remember the deceased with a bad word, his soul may suffer in the other world. It is better to remember something good about him. Let us recall the well-known proverb: only good (or nothing) can be said about the deceased. Signs have it that you can't speak during a minute of silence. At this time, it is forbidden not only to speak, but also to make any sounds in general. Going from place to place or digging through things is considered disrespectful to the deceased. You should not wear the clothes of the deceased. Someone believes that a person should not wear the clothes of the deceased, others, on the contrary, see this as an opportunity to touch a loved one, to be with him at least in this way. It all depends on feelings, personality and mood. Another thing is mortal clothing - the one that was on the person at the time of death. If the deceased was not immediately taken to the morgue, then the process of decomposition begins in his body. It is dangerous to wear or pass on clothing that was on the body at the time of death and it is best to get rid of it. The idea that the shoes of the dead person affect his soul was reflected in a number of beliefs. Such, for example, that you should not tie knots on the shoes of the deceased, since his soul will be tied in a knot, and he will appear in the dreams of the living people. What is the meaning of dreams about dead people? Dreams about our late loved ones should not be ignored as they usually carry tips and important information. What are the main funeral traditions and omens in Islam? Most of all, it is striking with what trepidation the relatives and friends of the deceased try to fulfill all the prescriptions of Sharia and bury a loved one like a real Muslim. Starting from the dying state and for a year after the funeral, relatives will diligently perform certain rituals. Many of them may seem unusual to a person not familiar with them, but for true Muslims, rituals are important and sacred. The Quran calls for preparing for death throughout life, so that at the end of it, one could accept such a difficult test with light heart. Special rituals prescribed in the Shariah begin to be performed while the sick person is still alive. The dying person is laid on his back with his feet to Mecca, which personifies the path of the soul to the holy place. It is necessary to help the sufferer cope with thirst by giving a sip of cold water. If possible, drip pomegranate juice or Zam-Zam - sacred water into his mouth. It is forbidden to cry loudly in Islam, so that the dying person can focus on his last test and not grieve for the worldly. Therefore, compassionate women may not be allowed to the bed or taken out of the house altogether. Immediately after death, the eyes of the deceased should be closed, his arms and legs straightened, and his chin tied. The body is covered with a cloth; a heavy object is placed on the stomach. A Muslim funeral should be held as soon as possible, preferably on the same day. Therefore, usually the followers of Islam are not taken to the morgue, but immediately prepared for burial. Islam has a strict attitude towards purity. If cleansing rituals are not observed, the body of the deceased is considered defiled, and the soul is not ready to meet with Allah. According to Muslim customs, men and women are buried barefoot, dressed in a simple shirt (kamisa) and wrapped in several pieces of linen. A wealthy and respected Muslim, who has not left any debts behind, is wrapped in expensive fabric. But not in silk: a Muslim man is forbidden to wear silk even during his lifetime. Muslim burials take place only in the cemetery. Cremation is strictly forbidden in Islam, it is equivalent to burning in hell. If a Muslim cremated the body of a relative, it means that he condemned his loved one to hellish torments. They lower the deceased into the grave with their feet down, while holding a veil over the women: even after death, no one should see her body. The Imam throws a handful of earth into the grave, recites the sura. Then the burial place is poured with water, the soil is thrown seven times. After the funeral of the Muslim, everyone leaves, but one person remains to read prayers for the soul of the deceased. A commemoration in Islam is, first of all, a commemoration of the deceased, a prayer for his soul and an opportunity for the family to unite in order to survive the grief more easily. Alcohol is strictly prohibited at the Muslims' commemoration. What are the funeral traditions and customs in Judaism? The Jewish worldview, based on the belief in the immortality of the soul and life in the next world, sees death in nothing more than a transition from life in the material world to life in the spiritual world, the world of absolute goodness. If a person is dying, Viduy is said at his bed - a prayer for the confession of sins, which is read on Yom Kippur, and if the dying person can, he repeats it as well. Immediately after death, they open all the windows in the house and pour out the water that was in any vessels. The body of the deceased is placed on the floor (legs together and hands on the chest), his eyes are closed and his face is covered with a white cloth. Candles are lit at the head of the bed, and one of the relatives or close friends sits down by the body to read the book of Tegilim. After the body of the deceased is cleansed, they wrap it in a shroud and cover it with a tallit, and the funeral procession sets off. The deceased should be buried as soon as possible. According to traditions, one cannot bury a Jewish person on Saturdays and holidays, and only outside Israel on the second day of the holidays is the burial performed. On Saturday or holiday the deceased person is buried at night, immediately after the end of Saturday or holiday. Before the body is lowered into the grave, the tallit is removed from it, in which it has been wrapped up until now, and the dead person is buried in a shroud. The shroud is the same for everyone - for in the world of truth, everyone is equal, “a nobleman is no different from a poor man” when he comes to the Creator for judgment. Taking part in the burial is an important commandment, and during the funeral everyone tries to carry the funeral stretcher at least a little and participate in filling the grave with soil. Participation in a funeral is called the "true good" that is given to the dead. When the participants in the funeral procession return from the cemetery, they should wash their hands: three times for each hand. After washing your hands, it is not customary to wipe them – this is a tradition to prolong the memory of what happened. When leaving the cemetery, the participants in the funeral stop and form two rows, and the mourning men pass between them taking off their shoes. From the cemetery, the mourners come to the house where the deceased lived. There they will sit shiva - seven days of mourning. According to superstitions and omens, mirrors and pictures depicting human faces should be covered (it is also customary to cover the TV screen, on which faces appear, too) in the house where the mourners observe mourning. Those who come to the mourners do not greet them, and when they leave, they do not say good bye. It is also not customary to shake their hands. By tradition, all the food of this meal should have a round shape: bagels, lentils, eggs - a hint that our whole life is a “revolving circle”, like a carousel, which revolves back to the beginning of the circle. So the present sadness will pass and the days of fun will come again... What are the customs and traditions of funeral in Hinduism? In modern India, Hindus make up about 80% of the population (Muslims make up about 15%, Christians - about 2.5%), so we will tell you about Hindu rituals. Traditional Hindu funerals involve the burning of the deceased (exceptions are made for children under 2 years old, saints and some other special categories) - this supposedly should help the soul to free itself from the mortal shell and provide it with a favorable rebirth in a new body. From time immemorial, Hindus have burned their dead on funeral pyres, using wood as fuel. The ceremony is led by a representative of the brahmana caste, who may also advise adding certain additional ritual objects to the fire, which should make the ceremony more successful and increase the chance that the deceased will be lucky in the future life. Cremation usually occurs within 24 hours of death, during this time the deceased should (preferably) remain at home. For the ceremony itself, people are dressed in white (black clothes are considered unacceptable at Hindu funerals), but no special clothing is provided. Under the guidance of a priest, the congregation sing hymns and mantras, after which they make offerings to the gods. It is permissible to bring flowers to a Hindu funeral, but not food. The fruit is brought to the farewell ceremony, which is usually held 10 to 30 days after the funeral. The body is laid on a fire in an open coffin so that all those present can see it until the moment of burning, but touching the body is highly undesirable (Hindus believe that this defiles the one who touches it). It is also considered indecent to use recording devices during funerals. After the funeral, the ashes of the deceased are collected in an urn and scattered. It is preferable to do this over the Ganges River. After that, it is assumed that the clergyman will monthly pray for the soul of the departed, and a year after the death, it is expected of the family of the deceased that they will arrange a large celebration. Representatives of the lower classes (who are the absolute majority) are often forced to resort to loans in order to provide their loved ones with the lowest level of rituals. The fear of being branded as “inattentive for the dead” in society is stronger for them than the fear of falling into debt. Electric and gas crematoria in India are currently a rarity: there are only a couple of dozen of them, and so far they exist only in large cities. Their spread is hindered by religious considerations: many traditional rituals simply cannot be performed if the body is burned in a crematorium. When a Hindu dies, the family contacts the priest. Those present recite mantras or turn on mantra recordings. When death is near, the body is transferred to a grass mat on the floor. A small amount of water from the river Ganges is put into the mouth of the dying person. If this is not possible before death, then these actions should be performed immediately after death. As soon as a person dies, the audience avoids touching the body, since at that moment it is considered unclean. The preparations for the funeral begin immediately. The funeral should take place, according to tradition, as soon as possible by the next sunset or dawn, whichever comes first. A priest he can help in the decision-making process and guide the family to a Hindu funeral home. Organ donation is acceptable to Hindus as there are no Hindu laws prohibiting the donation of organs or tissues from a deceased person. Embalming is also acceptable in Hinduism. Traditionally, in India, the body of the deceased is washed by the family and loved ones. For ritual ablution, the head of the deceased must be directed to the south. Once the body is cleansed, the big toes should be tied together and the hands are placed with the palm in the prayer position, the body should be wrapped in a simple white sheet. If the deceased was a married woman who died before her husband, she is dressed in a red robe. Upon returning home, all family members wash and change clothes. Then the family gathers at the table. The priest, at the request of the relatives of the deceased, can visit the family and cleanse the house with incense. The cremation of the deceased marks the beginning of the mourning period, which lasts 13 days. During this time, the family of the deceased must remain at home and receive guests, but some funeral rituals may differ from community to community. The photograph of the deceased can be displayed in a conspicuous place and decorated with a garland of flowers. Unusual funeral rituals and traditions from all over the world. China: a bride for the deceased and a stripper. For 2000 years in China there is a tradition - to find a bride for an unmarried deceased man. Chinese relatives believe that he will suffer from loneliness in the afterlife. The ritual was ubiquitous until the 20th century, when communist officials decided to ban Yin Huan as part of harmful superstitions. But even now, there are sometimes incidents of the theft of dead women. Also in China, the number of guests at the funeral is considered important - this reflects the social status of the deceased. A huge crowd must come. Since the 90s, the tradition has come into vogue to order strippers for funerals, who should beckoning people to burial. It doesn't matter that those who came are completely unfamiliar with the deceased. Japan: part of the ashes - for the family, part of it - for corporations. In Japan, due to the small amount of land, it is customary to cremate the dead, and not to bury them in coffins. Since the Japanese spend most of their lives at work, and almost do not take days off and sick leave, the work team becomes their second family, like a corporation. Therefore, after cremation, the ashes are divided into two parts: one is buried by relatives in the ancestral grave, and the other - in a corporate cemetery. Monks from the Japanese province of Yamagita practiced early preparation for mummification in 3 stages from the 11th to the 19th centuries: they reduced fat mass, ate trees, drank poisonous juice, and lay in a grave with a bell to wait for death. This was considered a special spiritual practice, although this phenomenon is not fully substantiated by the scientific world. Ghana: creative coffins and dances. Many people have recently seen a short video of tough guys in handsome suits carrying a coffin, smiling and dancing. This tradition is practiced in Ghana. The funeral should be fun with dancing. This is how the completion of the earthly journey is celebrated. The whole community gathers money for the funeral, since the coffin and treats are very expensive – almost the annual earnings of an average Ghanaian. Coffins are made in interesting shapes, depending on what the deceased was doing: for a musician – there will be a coffin in the shape of a guitar or a violin; for a pastor - in the shape of a Bible; if the dead person was a seller - in the shape of a wallet; and for a fruit picker – the coffin will be in the shape of a pineapple. Philippines: having a drink with the deceased. In the Philippines, the deceased is exhibited in front of the house for several days. He is dressed in his best clothes, tied in a sitting position to an armchair; a cigarette and a glass of alcohol are inserted into the palm of his hand. People from the community can come to say goodbye to the deceased at any time, say the last words, parting words. When the temperature and animals begin to influence the body too strongly, then it is buried. In the highlands, coffins with deceased Filipinos are hung on the rocks. It is customary to prepare coffins on your own in advance. High-altitude funerals are held for reasons of saving of sown land, and relatives also believe that their loved ones look after them from a height. While traveling, itinerant fellow citizens also often encounter merry funerals, where there is dancing, drinking and unusual forms of commemoration. You can take part in this procession - some local residents are interested in a large number of parting people. Perhaps it will help people come to terms with the loss and move on more easily. How did omens and superstitions appear? Sergii Haranenko You are welcome to share your dream here Welcome to CheckMyDream Joining our website you accept Checkmydream's Privacy Policy
A biography of paul revere an american patriot and silversmith After a year in the Army, Revere returned home to Boston, took over the family silver shop, and married his first wife, Sarah Orne. Revere's extraordinary skill also extended to his carving picture frames for the painter John Singleton Copley — I told him I esteemed myself a man of truth, that he had stopped me on the highway, and made me a prisoner, I knew not by what right; I would tell him the truth; I was not afraid. paul revere american revolution Visit Website By all accounts, the young Revere was a serious and committed artisan. King George III had issued a demand for its retraction. Paul revere university location He was retained by the provincial congress as a courier, and he printed local currency which the congress used to pay the troops around Boston. This letter, adopted in response to the Townshend Acts , called for united colonial action against the acts. He boarded in Watertown , where he was eventually joined by Rachel and most of his children Paul Jr. Revere such information as will inable him to Conduct the business on his return home. This system was an improved version of an old network of widespread notification and fast deployment of local militia forces in times of emergency. Devens told me, and of my being stopped. Later Life Paul would serve in the American Army during the revolution. But it was his ride on April 18,that etched his name into history, especially after it was commemorated later by the poet Henry Wadsworth Longfellow. From this, he evolved a neoclassic style after the Revolution. Rated 5/10 based on 116 review Paul Revere Biography
Skip to main content Table 3 Level of knowledge about schistosomiasis in the two villages From: Elimination of schistosomiasis requires multifactorial diagnostics: evidence from high- and low-prevalence areas in the Nile Delta, Egypt Knowledge QuestionKnowledge level (Percentage) What is bilharziasis?27.746.015.69.656.744.4 Which organs are severely affected?8.613.021.026.660.460.4 How do people get infected?5.33.541.623.953.172.6 How does bilharzia infect via canal water?17.410.648.263.034.426.4 Which sources pollute the canal water?36.629.03.58.859.661.2 How does human excreta reach the canal?64.079.520.514.415.56.1 What, and where are snail habitat?40.555. What is the role of snails in the life cycle?58.968.92.82.638.328.5 How can one identify bilharziasis?21.833.831.328.547.037.7 What are the types of available treatment0. How can one be sure of cure after being treated?21.938.611.623.766.537.9 How can you protect water stream from Bilharzia?33.641. How can you protect yourself From Bilharzia? Overall Knowledge43.859.235.530.420.710.4 1. The total maximum score equaled 31: satisfactory (≥ 75%), Unsatisfactory (74–50%) and poor (< 50%)
Ask a techie on the best way to learn coding, and the immediate response will be, “Use Linux” to get hands-on experience in programming. Having read scores of articles reiterating on the rising popularity of Linux systems in academia and professional services, we were just as curious to understand why techies swear by Linux OS over the Windows OS? More so, linux systems have captivated the interest of coding professionals and IT engineers over the years. The glory attributed to this basic computing platform stays untarnished through time, despite many rivals leaving no stone unturned to outpace each other. Why Do Programmers Use Linux To Learn Coding? Linux or any other Unix program is more popular with programmers for a unique reason, it is referred to as a free-to-use platform for a community of developers. With lots of programming tools available for free, some distros have them pre-installed, while some others are sold as package repositories. Many parser generators, debuggers, editors and compilers are all available for free. Linux being an open source operating system, it simply means programmers can read the source code of the programs. It is undoubtedly one of the best ways to learn coding for starters. Building upon the Unix tradition, Linux tools do not need frequent upgradation like Microsoft OS demanding users to invest time for frequent software upgrades. Linux OS saves users of the hassle by offering tried-and-tested tools which can be implied even today on modern distribution networks. The classic Unix tools have seen improvisation over the years, but at the core nothing seems to have changed in the way Linux simplifies coding complex programs for programmers to write codes effortlessly. Ever imagine C or C++ saving you of the pain to read thousands of codes before you pen one? Serious Linux users abide by the command line to develop complex programs easily on the basis of a simple set of building blocks. Best Linux Distros For Programmers While Distros such as Ubuntu and Debian have managed to allure users to be chosen as the top picks, there are many other Linux-based operating systems for beginners to choose from as well. To name a few – openSUSE, Arch Linux, Fedora, CentOS, Gentoo, Raspbian and many more to explore. Debian GNU/Linux Is not recommended for those who are new to Linux OS and are learning the basics of programming. It is a rock-solid stable system packed with tons of open-source goodies for advanced programmers and system administrators. Is the most commonly used Linux distro for coding purposes. It’s debian-derivative is used in cloud and server applications. The LTS version of Ubuntu enjoys community support for 5 years with ensured stability of a fixed release. Owing to dislike expressed by many programmers for Unity desktop environment, it has made some recent changes to switch back to GNOME desktop. OpenSUSE Linux Distro The openSUSE Linux distro offers a stable environment for programming and is available in two versions – its LTS release the openSUSE leap remains up-to-date to offer guaranteed stability, whereas the Tumbleweed is the rolling release for those who love to explore and experiment working on new softwares every time. Red Hat, CentOS Distro Basically a community version of Red Hat, CentOS distro is a must-try Linux distribution for those willing to test the working of server on a desktop. Its highly stable system allows programmers to develop applications safely using Xen virtualisation and YUM for package management. Arch Linux This is an all-time Linux distribution favourite among Linux enthusiasts because of its easy customisation feature, the Arch Linux. The distro’s repository is always up-to-date and one-time installation is sufficient. On release of a newer better version every month, you get to work on an updated machine. So focus on programming right and entrust other worries to the Arch Linux community of experts to resolve trivial concerns. If you want to rebuild an entire system for specific CPU architecture, and possess sufficient money to invest into one, then Gentoo gives developers the flexibility, stability and power to code all they want. This program is highly recommended for experienced Linux users who wish to take on new challenges in developing Linux packages. If you’re an expert, try Gentoo. Raspbian, the Raspberry Pi mini-computer is largely used in schools today to teach the basics of coding and Linux programming. Since Raspbian OS comes loaded with many programming tools, useful ports and connector pins, it makes for a perfect Linux OS to learn coding. If you want to learn things the other way around using DIY approach, then the combination of Raspberry Pi and Raspbian will help you hack into the secrets of coding it right, always! If you want to get started writing programs and complex software codes, then pick a best Linux distribution system and start exploring tools that can help unleash your programming potential to the fullest.
Are BBQs Significantly Polluting Air in Poland? A Simple Comparison of Barbecues vs. Domestic Stoves and Boilers Emissions Artur Jerzy Badyda , Piotr Krawczyk , Jan S. Bihałowicz , Karolina Bralewska , W. Rogula-Kozłowska , G Majewski , Przemysław Oberbek , Andrzej Marciniak , Mariusz Rogulski The study attempts to compare the total annual emissions of selected air pollutants emitted during occasional grilling and the emission of the same pollutants from small domestic heating installations. For this purpose, in the absence of any data on the emission of pollutants during grilling processes, tests were carried out consisting of measuring the concentration of air pollutants in exhaust streams from two types of grills (solid fuel grill powered by charcoal briquette and gas grill powered by liquid propane), using popularly prepared dishes (previously marinated meat and raw, seasoned mixed vegetables). The concentrations of PM2.5, CH4, CO, CO2, H2O, NH3, N2O, NO, NO2, SO2 were measured in the exhaust stream from both grills using a particulate matter (PM) measuring device and a portable spectrometer, separately while grilling the same portions of meat and vegetables. Then, considering the available data on Poles’ barbecue habits, the emissions that are released into the air during occasional grilling were estimated. The calculated emissions were compared with the data on emissions from domestic heating installations used in Poland. It has been shown that during grilling, as much as 2.30, 92.07, 4.11, 3.83, 2.96, and 9.81 Gg of PM2.5, CO, NOx, SO2, NH3, and CH4 may be released into the atmosphere in Poland, respectively. In the case of PM, the amount of the pollutant emitted to the air is over 100 times lower than the emissions caused by the operation of small heating installations. In the case of other pollutants, the differences are smaller. Nevertheless, emissions from grills should not be underestimated as, in certain periods of the year, these sources may be responsible for not meeting the air quality standards in selected areas of the country, and thus the excessive exposure of people to pollutants resulting in negative health consequences. Therefore, attention was paid to the legitimacy of abandoning the use of charcoal and charcoal briquette grills and replacing them with gas-powered grills or electric ones, not only due to the health benefits of food and lower human exposure, but also by the reason of ecological values. Strona publikacji na serwerze wydawnictwa Publikacje pracowników w Bazie Wiedzy PW Zmiana rozmiaru fontu
Select Page Tesla’s Computer Plate Just like watching TV, with the electromagnetic fields disturbing our electrical systems, so does our laptop screens and computer screens. Ever noticed, how after a while of sitting and staring at a computer screen, that your eyes get sore, you get tired, irritated and tend to lose focus or clarity, jumping at snacks or coffee to keep you going! This is because, like a TV, there is an electromagnetic field that radiates out perpendicular from our computer and laptop screens in the same radius of the length of the screen, which pierces through our electrical energy barrier’s and causes stress in the body and messes with our neurological system. Along with this, we have Wi-Fi and 5G, which pierces even harder into our cellular membranes, causing all sorts of chaos in the system. We are not designed to be sitting in front of a computer screen 8 hours a day, yet many of us do and pay for it later. It is just so vital we protect our electrical systems and energy fields, as we know that the health of our energy reflects our thinking, our emotions, and thus what we manifest into our reality. What Do Tesla’s Computer Plates Do? Tesla’s Computer Plates effectively negate the effects of Electromagnetic Fields coming from Computer and Laptop Screens, VDU screens (Visual Display Units), as to keep our electrical energy undisturbed by the chaotic interference of EMF’s lingering in our energy fields. Tesla’s computer plates also alleviate the stress coming from Wi-Fi and gaming devises, keeping our electrical energy field strengthened and harmonious with such devices. When placed on top of the computer screen with the concave side facing down into the area just behind the screen, you should instantly notice the lack of eyestrain whilst using the computer or laptop. NOTE: We recommend using the Computer Plate together with the Electron Stabaliser to ensure that all electrical chords, chargers and currents running into the computer or laptop are neutralised, as to ensure your energy is not disrupted. Benefits of Tesla’s Computer Plate include: • Increases your focus and attention span • Clears your mind and allows for greater concentration • Stops computer eye strain • Increases productivity in the work place • Makes for a happier and more peaceful workplace • Adds life force into water (making it clearer and taste better) • Adds life force into food (adding sweetness and flavour) • Sweetens sour fruit • Brings to life unhealthy plants • Aids in the healing of chronic pain • Speeds up the healing process of injuries • Helps heal cuts, burns and bruises when placed on the affected area • Helps relieve head aches • Helps relive inflammation and arthritic pain • Increases productivity threefold Lastly, Tesla’s Computer Plate aids in the maturing of poor quality or cheap wine, which is a God Send for Wine Lovers!! Tesla’s Computer Plates have not only helped many sensitive individuals be able to tolerate their laptops for lengthened periods of time, without stress, which was a by-product of use in the past, but they have also managed to turn around entire work places, from once stressful and moody work colleagues, to a calm and productive environment, when a computer plate was provided at each work station. This alone has transformed businesses!! For Detailed Information on How Tesla’s Computer Plate & Tesla’s Innovational Technology Works, Download the Free Information Booklet HERE Aura Photography of Tesla’s Computer Plate The Pictures below are taken with an Aura Camera. The picture on the left below, shows an individual using a computer without a Computer Plate. The picture on the right, shows the same individual using a computer, just after applying the Tesla Computer Plate. You can see the difference in the colour of aura from a red, yellow, orange (stressed energy field) to a calm blue (peaceful undisturbed energy field). Using Computer Without Plate Using Computer With Plate How Does the Tesla Computer Plate Work? As with all Tesla Products, the Computer Plates “channel” through or act as a transceiver to specific photons (packets of light energy) coming from the Central Sun, and then transmute this energy from the middle of the plates (70% on the concave side, and 30% on the convex side). These specific photons “light energy” shine out in a radius of a metre, so if a frequency camera was used, you could see the light coming in from the edges of the plate and out again from the front and back of the plate and radiating a metre away in an oblong type shield/shape. These titanium Computer Plates, have had their vibrational frequency altered in a way, as to combat against Electromagnetic Fields from the Computer Screen, Laptops and Wi-Fi, being their specific duty in the Tesla family of Products. Tesla’s Computer Plates are the Bees Knees of Counteracting Wi-Fi, Electromagnetic Radiation (EMRs) and Electromagnetic Fields (EMFs) from work stations, laptops, gaming and computer screens, which is causing too much disharmony and thus dis-eases in the body, and is way too strong for the influx of gifted children who are extra sensitive, highly empath and intuitive, now coming into the world to teach us a lighter way of being! NOTE: No other company has access to this technology of Nikola Tesla’s Workings. Therefore Tesla’s Innnovational Products are the only product using the very technology designed by the 20th Century Genius, which we now have available at our finger tips, to counteract Electromagnetic Radiation (EMRs) and Electromagnetic Fields (EMFs) as well as Geopathic Stress. We thank Nikola Tesla for thinking ahead and passing on his knowledge to the next Century, so we may all benefit today! RRP = $152 AU per Computer Plate 10% DISCOUNT JWNK Online Price = $135 AU To Order your Computer Plate with 10% OFF CLICK HERE Postage: $15 Express Post Get Your FREE Juicing with Nadia K Information Booklet Your Cart Your cart is emptyReturn to Shop
Military Wiki Krzysztof Radziwiłł Born (1585-03-22)22 March 1585 Biržai, Polish-Lithuanian Commonwealth Died 19 November 1640(1640-11-19) (aged 55) Svėdasai, Polish-Lithuanian Commonwealth Religion Calvinist Spouse(s) Anna Kiszka Children with Anna Kiszka: Janusz Radziwiłł Katarzyna Radziwiłł Parents Krzysztof Mikołaj "the Thunderbolt" Radziwiłł Katarzyna Ostrogska Prince Krzysztof Radziwiłł (Christopher Radvila, Lithuanian language: Kristupas Radvila ) (22 March 1585, Biržai – 19 November 1640) was a Polish-Lithuanian noble (szlachcic), and a notable magnate, politician and military commander of his epoch. Sometimes referred to as Krzysztof Radziwiłł II, to distinguish him from his father, Krzysztof Mikołaj 'Piorun' Radziwłł. A proud magnate, sensitive to the point of ancestral honour and heritage, convinced of his vocation to play a leading role in the country. He was a successful military leader, but selfishness and pride divided his supporters and weakened his significance in the politics. As the leader of the Lithuanian Calvinists, he was often threatened and disliked by the zealous and ultra-Catholic King of Poland, Sigismund III Vasa.[1] Owner of Biržai, he was Field Hetman of Lithuania from 1615, Castellan of Vilnius from 1633, Voivode of Vilnius Voivodeship from 1633,.[2] Great Lithuanian Hetman from 1635,.[2] starost of Mahilyow, Bystrzycadisambiguation needed,[3] Žiežmariai and Seje. In 1632 he served as Marshal of the Sejm.[2] He successfully fought in Polish Livonia against Sweden from 1600s to 1617. He took part in the campaign against the Swedes in the Baltic region in 1621–1622 (Polish-Swedish War), where he agreed to a controversial truce (as he had no permission to negotiate it from the king or the Sejm).[2] Important commander of Lithuanian forces during the Smolensk War, where he was instrumental in obtaining the capitulation of the Russian army in the relief of Smolensk.[2] Afterwards, in 1635, he retired from his military career, concentrating on administering his estates. A Calvinist, he was a great protector of Protestants in Lithuania. He was an opponent of Catholic king Sigismund III Vasa, but a supporter of his more tolerant son, Władysław IV.[2] An advocate of Władysław's marriage to a Protestant princess, he distanced himself from the king after Władysław declined this marriage proposal. He opposed an alliance between the Commonwealth and the Habsburgs. On his lands in Kėdainiai he founded a major Calvinist cultural and religious centre,.[2] which flourished till the 19th century as a center of the Polish Reformed Church. His large horse stables were famous throughout Europe.
November 28, 2021 Global Warming and the Hypocrisy of the Rich Countries Prabir Purkayastha FIGHTING global warming is not just providing a path to net-zero carbon emissions for all countries. It is also about how to meet the energy needs of the people while doing so. If fossil fuels are to be given up as they need to be, countries in Africa and a significant part of Asia, including India, need an alternate path for providing electricity to its people. What is the course open to the poor countries if they do not use the fossil fuel route that the rich countries have used? And how much will such a route cost, and who will pay the bill? This discussion was completely absent from the COP26 agenda as the financing of a low carbon emission path was delinked from cutting down carbon emissions and is now scheduled for next year. Some numbers here are important. The European Union plus UK (EU+) produce more than twice the carbon emissions than the entire continent of Africa, with less than half of its populationWith less than a quarter of India’s population, the US emits significantly more carbon emissions than India. A simple argument regarding replacing fossil fuels with renewables without bothering about funding, is the cost of electricity from renewables. If the cost of electricity using renewables is lower than that from fossil fuels, it should be possible to phase out fossil fuels completely with renewables, and we do not need to bother about funding. Or so the argument goes. While the cost per unit of electricity with renewables is today lower than with fossil fuels, we need to build three or four times the capacity from renewables to supply the same amount of energy that we would get from fossil fuel plants. That means we will have to invest four times the capital in renewables to generate the same amount of electricity. We need much more capital even if the cost of electricity is the same as from fossil fuels, as we do not have to pay for the “fuel”: nature provides that free. If we are a rich country, that may not be a problem. But for a poor country trying to build its basic infrastructure of electricity, roads, railways, other public infrastructure, including schools, universities and health, it is not easy without financial support from the rich countries. This is why the rich countries asking the poor countries to give net-zero pledge without committing any money is completely hypocritical. Tomorrow, they can, and most probably will, turn around and say that you made the commitment, now borrow from us at high rates of interest and make good on your promises. Or face sanctions. In other words, a new form of green imperialism. The second problem with renewables as the main source of electricity is that there are significant additional costs that the grid will have to undertake in terms of short-term or long-term storage of electricity. This is to balance the short term or the seasonal fluctuations that may arise. For example, this year, Germany saw a significant stilling of winds in summer, leading to a sharp fall in wind generation. In their case, they balanced it with an increase of electricity from coal-fired plants, with their greenhouse gas emissions going up significantly. In a scenario that such plants do not exist, what will countries do? While daily fluctuations can be met with large, grid sized batteries, this is not feasible for seasonal variations. They will have to either use pumped storage schemes with hydroelectric power or store hydrogen in large quantities for use in fuel cells. A pumped storage hydroelectric scheme means pumping water up to a reservoir when there is a surplus of power on the grid and using it to produce electricity when there is a shortfall. Storing hydrogen in quantities large enough to meet seasonal grid requirements is still another idea whose technical and economic feasibility needs to be explored before it becomes a real choice. The point here is shifting to a grid entirely based on renewable energy is still technologically some distance away. And we may need to use concentrated sources of energy – fossil or nuclear – to meet the requirement of daily or seasonal fluctuations. This means developing new technologies. In fossil fuels, it means using carbon capture, that is, not letting carbon dioxide escape into the atmosphere pumping it into underground reservoirs. All such carbon capture projects in rich countries were given up in the belief that renewables will solve the problem of carbon emissions. It is now clear that renewables as the only source of energy in a grid are not enough, and we may need to look for other solutions as well, including carbon capture. In the short-term, nuclear power does not appear to be a solution. This means gas, oil, and coal are the only short-term solutions before us. And here, the duplicity of the rich countries becomes clear. Europe and the US have enough gas resources. India and China do not; they instead have coal resources. Instead of arguing how much should each country emit greenhouse gases, they decided to focus on what fuel needs to be phased out. Yes, coal emits twice the amount of carbon dioxide compared to fuel gas for producing the same amount of energy. But if you produce twice the amount of energy from fuel gas as from coal, you will still produce the same amount of carbon emission. If the US or the EU+ (EU plus UK) is producing more carbon emissions than India or Africa, why ask phasing out only coal, while no such targets are taken by the US or EU+ for phasing out their carbon emissions using gas? This is where the energy justice issue becomes important. The US per capita energy use is nine times that of India, the UK’s six times. If we consider countries in sub-Saharan Africa such as Uganda or the Central African Republic, their energy consumption is even lower, i.e., US consumes 90 times, UK 60 times more than these countries! Why should we then talk of only which fuels need to be phased out and not how much emissions that countries need to cut and cut immediately? I am not raising here the issue of an equitable share of the carbon space, and if a country has used more than its fair share, on how it should compensate the poorer countries. I am simply pointing out that by talking about net-zero and phasing out of certain fuels, the rich countries are continuing on their path of excess carbon emissions while changing the goalposts continuously for others. The last word on hypocrisy is Norway’s. ]At a time that Norway is expanding its own oil and gas production, along with seven other Nordic and Baltic countries, Norway has been lobbying with the World Bank to stop all financing of natural gas projects in Africa and elsewhere (Foreign Policy: Rich Countries’ Climate Policies Are Colonialism in Green, Vijay Ramchandran, November 3, 2021). While Norway may have been the most blatant, 20 countries combined with Norway to move similar resolutions in COP26. For them, climate change negotiations is how to keep their dominant energy positions while denying not only climate reparations but even finances for the poorest of countries that are trying to provide their people with subsistence energy. It is clear that we have no future if we do not stop the continued emission of greenhouse gases. But if we do not also find a path for the poorer countries to meet their minimum level of energy needs, we will also see the collapse of huge swathes of these countries. Do we think we can live with countries in sub-Saharan Africa living on 1/90th the energy consumption of the US without the consequences for all of us? And Modi and his followers may believe that we are on the way to becoming a developed country, even a superpower. The fact is, in energy terms, we are, in fact, closer to Africa than we are to China or the club of the rich: the US, EU and the UK.
Skip to main content How “forever chemicals” might impair the immune system By April 13, 2021April 16th, 2021PFAS in the news Stain-resistant carpets and nonstick pots were once the epitome of “better living through chemistry,” their space-age properties conferred by molecules known as perfluoroalkyl and polyfluoroalkyl substances (PFAS). But in the early 2000s, researchers began to discover that PFAS were somehow reaching the farthest corners of the planet—from polar bears in Alaska (1) to pilot whales in the Faroe Islands of the North Atlantic (2). These molecules contain chains of carbon peppered with fluorine atoms, which together form one of the strongest known chemical bonds. That helps these chemicals excel at repelling grease and water but also makes them astonishingly resistant to degradation in the environment (3). Amid a flurry of new studies, scientists are still figuring out what risks these ubiquitous “forever chemicals” pose to public health (see “PFAS Politics”). Epidemiologists and toxicologists point to myriad possible consequences, including thyroid disease, liver damage, and kidney and testicular cancers (4). Impacts on the immune system are a particular concern. Read more…
Advertiser Disclosure Our readers come first Technical vs. Fundamental Analysis: A Beginner’s Guide Technical and fundamental analysis are the two major strategies when it comes to researching and forecasting stock markets. Technical and fundamental analyses are techniques employed by investors and traders to analyze the current situation and prospects of different financial instruments. These two approaches analyze different variables to determine if the financial asset could experience a positive fluctuation in the future. They can also point to a negative outlook in some cases. Understanding these two techniques will help amateur investors in enriching their investment skills and, upon reading this article, you will know the basics of how both of these approaches work and how you can apply them to analyze different financial instruments. Fundamental Analysis Technical Analysis Definition Calculates stock value using economic factors known as fundamentals Uses security price movement to predict future price movements Data gathered from Financial statements Charts Stock bought When price falls below intrinsic value When trader believes they can sell it for a higher price Time horizon Long-term approach Short-term approach Function Investing Trading Concepts used Qualitative Fundamentals (e.g. business model, etc.) and Quantitative Fundamentals (e.g PE ratio, PEG ratio, EPS, etc.) Price trends, chart patterns, oscillators, volume and momentum indicators, moving averages, support and resistance levels Vision Looks backward as well as forward Looks backward Source: Diffen What Is Fundamental Analysis? The goal of fundamental analysis is to determine the intrinsic value of a certain financial instrument upon assessing certain qualitative and quantitative factors. For stocks and bonds, most of the information used to perform a fundamental analysis comes from the issuer’s financial statements. However, other sources, including industry publications and market research reports, may also be used to assess the competitive situation of the company and its prospects. On the other hand, for bonds, fundamental analysis can also be performed by using certain basic formulas to determine the fair value of the asset upon assessing its future cash flows and the creditworthiness of the issuer — among other factors. Benjamin Graham is considered the pioneer of fundamental analysis. This Wall Street legend wrote the book called “Security Analysis” back in 1934. The techniques he outlined back then for valuing financial instruments remain the foundation of today’s practices. Pro Tip: In a nutshell, fundamental analysis is a startegy used by investors to determine a stock’s real or “fair market” value. Fundamental Analysis Tools Both quantitative and qualitative variables are analyzed when trying to assess the value of a financial instrument using fundamental analysis. The following is a short list of the most common ones. Qualitative Fundamentals Qualitative fundamentals refer to the aspects of the business that cannot be quantified but that still play a role in shaping their financial performance over time. Business Model The business model of a company refers to how it generates money. It explains who its top suppliers and customers are, which channels it uses to distribute its products and services, and how much it charges. There are many different business models including retail, wholesale, subscription, franchise, software-as-a-service (SaaS), etc.  Understanding the business model of a firm is crucial to determining how it has made money in the past and how it plans to keep producing revenues in the future. Competitive Advantage A competitive advantage is a unique aspect of the firm’s value proposition that is not offered by its rivals. It sets the business apart from the competition and makes it more appealing to its customers. The more competitive advantages a business has, the better. Moreover, if those advantages are difficult to replicate, the business will be more valuable for investors as it will have what is known as “a competitive moat”. The leadership team of a company is essential for the adequate execution and achievement of its long-term initiatives and goals. For this reason, analysts pay close attention to who makes up this important team. Analysts typically looks into the management’s track record in previous jobs and what they have done for the company since they came on board. Another great way to analyze the senior management’s performance is to look at their promises and claims from five to 10 years ago to see if they have lived up to these expectations. Economic Factors All businesses depend and perform to some extent based on how macroeconomic factors of their respective domestic economies evolve. For this reason, a fundamental analysis also takes a look at the current economic conditions to assess the business’s valuation.  Some of the most important variables that analysts keep an eye on include prevailing interest rates, gross domestic product (GDP) growth, and inflation. These, among other factors, could have a sizable effect on a company’s performance in the future. Quantitative Fundamentals Financial statements and operating metrics are typically interpreted by analysts to determine the fair value of a company. These are some of the most common metrics that are typically used to assign companies a price tag. Balance Sheet The balance sheet of a business portraits its current financial condition. It includes a detailed breakdown of its assets, liabilities, and equity. A business with an elevated level of debt is perceived as riskier from a fundamental standpoint. On the other hand, the composition of a firm’s assets also plays a role in shaping its valuation.  For example, companies that are top-heavy in their balance sheet are considered more liquid and less risky than those that have a large fixed asset base as the latter group will be more sensitive to suffering impairment charges. Apple Balance Sheet Source: Marketwatch showing $AAPL balance sheet Income Statement The income statement of a company reflects its financial performance during a certain period — typically a quarter, a semester, or a fiscal year. This statement breaks down how much the company generated in sales during that time frame and progressively deducts the different expenditures that occurred during the period. This statement is typically analyzed to determine a firm’s profit-generation capacity while it could also be compared historically to analyze trends and sudden changes in the composition of the firm’s cost structure. The more predictable a company’s performance has been, the easier it should be to forecast how it should perform in the future.  Cash Flow Statement  The cash flow statement shows how much of the company’s earnings have turned into an increase in its cash balance. This statement adds back all non-cash expenditures and deducts all cash payments that affect the balance sheet to illustrate if the company has produced or burned its cash reserves during a certain period. For analysts, a company’s cash flow generation capacity is a great indicator of its financial profitability. One popular metric that is often used when performing fundamental analysis is the free cash flow metric.  This figure deducts essential capital expenditures from the firm’s net operating cash flow to understand how much money the company is producing to finance future investments or to distribute some of its profits to its shareholders. Price-to-Earnings (P/E) Ratio The price-to-earnings ratio uses data that can be easily obtained to estimate how many years it would take an investor to fully recuperate his/her investment in a stock if earnings remain unchanged. The higher the P/E ratio, the longer time it would take the investor to get his money back. As a rule of thumb, low P/E ratios (below 25) tend to be associated with potentially undervalued businesses while a high P/E ratio (above 25) tends to be associated with overvalued companies. However, other factors also play a role in shaping the trading multiples of a business aside from earnings alone. The price-to-earnings-to-growth (PEG) ratio was coined by the famous investor Peter Lynch. It seeks to smooth the inherent flaws of the P/E ratio by incorporating the company’s historical or forecasted earnings growth into the mix. In summary, a PEG ratio below one signals that a company is fundamentally undervalued while a PEG ratio above one indicates that the firm is overvalued based on its past or future earnings growth. Meanwhile, a ratio equal to one indicates that the business is fairly valued. Debt-to-Equity (D/E) Ratio Aside from earnings alone, fundamental analysis also focuses on certain components of the balance sheet — the most important being the firm’s debt. The debt-to-equity (D/E) ratio measures how leveraged a business is by expressing how many dollars of debt it holds per every dollar invested by its shareholders. A conservative D/E ratio would usually be below one, meaning that there is $1 in debt for every $1 invested by shareholders. Lower ratios may have a positive impact on the firm’s valuation as they would make the company more stable even under the most challenging economic scenarios. That said, certain businesses — i.e., capital intensive companies — will always have high D/E ratios due to the nature of their operations. For this reason, another way to assess if a business is highly indebted or not is to compare its D/E ratio with the average of its industry or peers. Profit Margins & Return on Equity (ROE) Ratio A company is only as valuable as its ability to generate profits for its shareholders. Therefore, the higher its profit margins, the higher the valuation that a business will obtain from prospective investors. In this regard, there are multiple profit margins that investors tend to analyze to assess a business’s value from a fundamental standpoint. These include: • Gross margin • Operating margin • EBITDA & EBIT margins • Net margins There are also adjusted versions of these margins that deduct non-recurring or non-cash items such as shareholder compensation and losses from discontinued businesses.  Moreover, investors can also assess the profitability of a company by measuring the percentage that its net income produces for every dollar invested by shareholders. This ratio is known as the Return on Equity (ROE) ratio and the higher it is, the more profitable the business is.  Fundamental Analysis Pros • It helps investors in determining the fair price (intrinsic value) of a business. • It provides an objective measure that can be used to determine if market prices are attractive or not. • It incorporates both qualitative and quantitative variables. • It can be used for multiple financial assets. Fundamental Analysis Cons • It will require some guesswork to estimate the future path of a company’s profits. • The past performance of a business does not guarantee future results. • It requires extensive accounting and corporate finance knowledge, especially if the business is large and complex. What Is Technical Analysis? Technical analysis is a technique that employs price and volume data to determine if the price of a financial instrument is poised to go up or down. The tools used for performing this kind of analysis include chart patterns, price or volume-derived indicators, and trend analysis. Technical analysts can interpret data from different time frames going from multi-minute or multi-hour (for day traders) to multi-day or multi-week (swing traders). Pro Tip: In a nutshell, technical analysis is a resource used by traders to research and evaluate investment opportunities utilizing price trends and patterns seen on charts. Technical Analysis Tools The toolbox of a technical analyst is quite extensive and even the most sophisticated traders come up with indicators to try to assess the direction that the price of a security will take in the future. The following is a summary of the most common ones. A trend indicates the direction that the price or volume of a security has taken in the past and it is used to predict the direction that it may continue to follow in the future.  An uptrend refers to a price or volume chart that shows a series of subsequent higher highs. Meanwhile, a downtrend is a chart formation that displays a series of subsequent lower lows in the price. Most traders prefer to “trade with the trend,” which means that they would rather follow the overall direction of the price or volume action rather than go against it.  Volume and Momentum Indicators Volume indicators use this data to determine if a stock is experiencing positive momentum. High volumes during positive sessions tend to signal that bulls have taken over the price action while high volumes during negative sessions tend to signal the opposite.  Volume indicators such as the volume profile can also be used to assess which prices have seen the highest trading volumes. These prices are typically highly contested and a break above or below them could be interpreted as a buy or sell signal. Meanwhile, momentum indicators are designed using historical prices (moving averages or exponential averages) to determine when the current directional trend of a stock is accelerating. Moving Averages Moving averages collect information from the past to possibly predict the future of either the price or volume of a financial instrument. There are simple and exponential moving averages. The former is just the sum of the prices of the last however many days divided into that same number while the latter puts more emphasis on the most recent data as it is considered more relevant. The most common moving averages used by traders are the seven-day, 20-day, 50-day, 100-day, and 200-day simple and exponential moving averages. Marketwatch chart showing $AAPL indicators Source: Marketwatch chart showing $AAPL stock indicators Support and Resistance Levels Support and resistance areas are price levels identified within a chart that show strong selling areas (resistances) and strong buying areas (support).  A support area is one where investors become highly interested in the instrument and buy it accordingly. Meanwhile, a resistance area is a price level at which investors are willing to sell or short-sell the asset as they anticipate an upcoming decline in its price. Oscillators are technical indicators whose value is derived from past price or volume data. They are designed to provide buy or sell signals to traders. The Relative Strength Index (RSI) and the MACD are among the most frequently used oscillators. The RSI, for example, tells traders if a stock is overbought or oversold based on its latest price action. In theory, an overbought stock is poised to decline as market participants have pushed the price too high too fast while the opposite can be interpreted if the oscillator enters the oversold area. Chart Patterns A pattern is a formation that can be identified in a price chart. There are simple patterns such as price channels and more complex setups such as the head and shoulders or cup and handle patterns. These patterns were designed to easily interpret what the price of an asset is doing within a certain time frame to possibly predict the direction that it will take in the future. One example is the head and shoulders pattern, which consists of a first high (first shoulder), a subsequent higher high (head), and a lower high (second shoulder). The pattern also identifies a neckline, which is a support area from which the price has bounced throughout the process of posting these three key highs.  In practice, this pattern tells the trader that bulls have exhausted their ammo upon pushing the price to the highest high (the head) and now bears are taking over – hence the second shoulder. Once the neckline has been broken, the pattern is considered confirmed and a sharp decline should follow as the bulls should capitulate. Technical Analysis Pros • It can be used to determine the best entry and exit prices for a particular instrument. • It illustrates, to some extent, the market’s sentiment toward the instrument. • It uses objective data (price and volume) to predict the future directional path of the asset. Technical Analysis Cons • It can be complicated to interpret unless the analyst has been appropriately trained. • Regardless of their complexity, no indicator or pattern is infallible. • It does not take into account the fundamental aspects of the underlying asset. What Are the Differences Between Technical and Fundamental Analysis? • Technical analysis focuses on price and volume data while fundamental analysis relies on financial statements and market research reports. • Technical analysis is typically used for short-term trading strategies while fundamental analysis is mostly employed by long-term investors. • Fundamental analysis seeks to determine the fair value of a business while technical analysis aims to determine the future direction that the price of an asset may take. • Technical analysis focuses on past data while fundamental analysis also incorporates an assessment of future projections. Who Is Technical Analysis Best For? Technical analysis is typically used by traders and investors who have relatively short investment horizons — often known as swing traders or tactical investors. That said, long-term investors could rely on technical analysis to identify the best time to enter a stock — for example, if it has recently bounced off a key support area. Who Is Fundamental Analysis Best For? Fundamental analysis is better suited for investors with a mid to long-term investment horizon. However, determining if a business is undervalued at a given point does not guarantee that the market will rush to correct its mistake — or that such assessment is necessarily right.  Therefore, long-term investors who use fundamental analysis must be patient to wait until the market either validates or rejects their thesis. That could take months, years, or, in some rare cases, decades. Final Thoughts Even though investors and traders have discussed for decades which of these two approaches is better, as with many things in life, neither is necessarily better than the other — they are just different.  Depending on what role you plan to take when participating in the financial markets — trader or investor — you may pick among these two techniques accordingly or you may use them in combination to strengthen your investment thesis. Stock Market Playbook Plus our best money tips delivered straight to your inbox.
Why is the place where the visible screen (on load) ends called the Fold? I have a hard time even writing out the definition for it, but when someone says The Fold, I know exactly what they mean. Why is it called the fold, or above the fold? Is the term appropriate to use when working with non-developers? • 46 With this question, I am an old man. – dotancohen Oct 17 '14 at 8:26 • 3 The fold isn't in the same place for every user though, it depends on the size of their window, right? – Michael Oct 17 '14 at 21:13 • @Michael: Generally, a web site should be designed so that most users will be able to see enough without scrolling to know that the balance of the page is likely to be of interest, and so that uses who want to use common navigation links will not have to scroll to find them. It doesn't matter if many users get to see a lot of information beyond the bare minimum; what's important is that even users with small-but-not-unusually-small screens don't have to scroll excessively. – supercat Oct 19 '14 at 21:10 • 3 It should also be noted that 'the fold'--as being an issue on the web--is a myth: boxesandarrows.com/blasting-the-myth-of-the-fold cxpartners.co.uk/cxblog/… uxmyths.com/post/654047943/myth-people-dont-scroll – DA01 Oct 20 '14 at 1:30 • So, in other words, it's not a relative term at all for UX on the web. – DA01 Oct 20 '14 at 1:31 It's from print newspapers; back in the day when broadsheets were more common, they were usually presented folded in half vertically, so the most important part of the front page was the portion "above the fold", which is the first thing most people see when they see the newspaper. Analogously, this is the first part of the website you see when a page loads, before the user unfolds the newspaper or scrolls down. With non-developers, it depends on their experience with jargon. Designers would generally understand what you mean, but the average layperson might not, since there's no clear foldable portion of a website or a screen. • 2 Indeed, the average lay person might not already know this jargon, but it is an appropriate term to use with them. You simply need to explain it first. Fortunately, the metaphor is simple enough that most people will get it straight away. Oct 16 '14 at 18:33 • 1 Please site sources. FineDictionary.com's Idioms for Above corroborates this etymology: "Above the fold - If a news story is important, it will be above the fold- in the top half of the page of a newspaper." Oct 16 '14 at 19:11 • 6 @3nafish: Sorry, I'm not super familiar with the rules still... I answered out of my personal general knowledge, is it expected that I should normally back that up with external sources also? – rach oune Oct 16 '14 at 19:30 • 6 It has nothing to do with delivery. When delivered, newspapers were/are usually rolled up. Newspaper vending machines (news racks) displayed folded newspapers. The portion above the fold was very important to get people interested in buying the paper from the vending machine. Photo: en.wikipedia.org/wiki/Newspaper_vending_machine#mediaviewer/… Oct 16 '14 at 22:17 • 1 @StephenOstermiller: Sorry, you're absolutely right. I should have said "presented". I'll edit my answer to reflect that. Newspapers in my neighbourhood are usually delivered folded but I shouldn't have assumed that's standard across areas. – rach oune Oct 16 '14 at 23:00 It comes from newspapers which are folded in half. Above the fold refers to content that is visible without unfolding or turning the newspaper over to see the 2nd half. This term was adapted to websites and their content that is visible without scrolling. Here is a picture of a newspaper. enter image description here Everything you can see is above the fold. The history of the term "Above the fold" comes from newspapers where the articles at the top were most visible when the newspaper was folded. To quote this Wikipedia article Above the fold is the upper half of the front page of a newspaper where an important news story or photograph is often located. Papers are often displayed to customers folded so that only the top half of the front page is visible. Thus, an item that is "above the fold" may be one that the editors feel will entice people to buy the paper. Alternatively, it reflects a decision, on the part of the editors, that the article is one of the day's most important. By extension, the space above the fold is also preferred by advertisers, since it is the most prominent and visible even when the newspaper is on stands. With regards to developers, while most developers I work with understand the above the fold term, if users or stakeholders dont understand, I explain it as the content that is visible when you first load a page and what you can see without scrolling. enter image description here However do note, above the fold is a relative term as it depends upon the resolution and the form factor of the device used to view the content. I also recommend reading this witty article about the fold • 3 A site should use what's above the fold to convince a reader that there's something good below the fold. To use the newspaper analogy, if nothing above the fold was of interest, nobody would read the paper; if everything of interest was above the fold, many people would read the paper, but nobody would buy it. Once people are drawn to a page, they'll proceed further, but if the page doesn't draw their attention quickly they're likely to leave and never return unless they've been told elsewhere what to expect. – supercat Oct 18 '14 at 4:07 Your Answer
Boster Pathways ➔ Neuroscience Alzheimer's Disease Pathway Alzheimer’s disease (AD) is characterized by an irreversible decline in cognitive and behavioral abilities that will eventually render the person completely reliant on others for basic activities in daily life. View pathway Kumar reviewed the pathophysiology and management of Alzheimer’s disease. In the paper, the development and promising therapies were discussed. Although there is enough information about the disorder, only limited choices exist for its management (Kumar, 2015). AChEls and Memantine, which are medications for its treatment, only treat symptoms and neglect the underlying factors responsible for the disease. However, the administration of vaccine seems positive because patients with early stages show improvement in their behaviour after Ab vaccination (Kumar, 2015). New solutions under verification include NOS modulation, and DNA vaccination (Kumar, 2015). Also, Mizuno constructed an Alzheimer’s disease pathway named “AlzPathway” which is accessible by everyone. The pathway is a directory of Alzheimer’s disease signaling pathway. It is the alpha and omega for decoding pathogenic of Alzheimer’s disease (Satoshi, 2015). The pathway is made up of 1347 molecules, 1070 reactions and 129 phenotypes in neuron, brain blood barrier, astrocyte synaptic, microglial cells and their positions. Brief History In 1901, Alois Alzheimer, a German psychiatrist, recognized the first fifty-year-old woman with Alzheimer’s disease. After five years, eleven more cases were described in medical journals (Alzheimer, 1907). Back then, old age was regarded as a risk factor of increasing dementia. Emil kraepelin was the first to explain it as a typical disease after debunking some information in the medical literature of Auguste D (Berrios, 1990). Alzheimer’s disease also called presenile dementia is a category of senile dementia described by Kraeplin in one of his books (Kraepelin, 2007). In the 20th century, Alzheimer’s disease diagnosis was restricted to those within the susceptible age group (45 to 65 years old) showing signs of dementia. After 1977, the concept was modified when a seminar on Alzheimer’s disease concluded that similar symptoms exist in senile and presenile dementia. Hence, this brought about the diagnosis of Alzheimer’s disease irrespective of age (Boller, 1998). AD Causes, Stages, Symptoms, Treatments Alzheimer’s Disease Causes Alzheimer’s disease (AD) is said to be the result of unusual accumulation of proteins in the brain tissues. Among them is amyloid which accumulates to build up plaques in the brain tissues. Another type of protein involved is TAU which accumulates to form tangles in the brain tissues. Though the root of this process is still a mystery, the fact remains that it exists several years before the symptoms become visible. Dysfunction of the brain tissues is also accompanied by reduction neurotransmitters that transmit information in the brain. Subjects with Alzheimer’s disease have low levels of acetylcholine (a neurotransmitter) in their brain. Subsequently, various sections of the brain shrivel especially the part that is in charge of memories. In special cases of Alzheimer’s disease, there is dysfunction of the brain cells and early symptoms include loss of vision or speech impairment. Alzheimer's Disease Stages There are four stages of clinical classification of AD described below. • Preclinical AD This phase is accompanied by mild symptoms which are usually ignored. It is typically regarded as mild cognitive deterioration. At this stage, the first pathological symptoms appear, strike the entorhinal cortex, and subsequently the hippocampus. Symptoms include mild amnesia that don’t affect day-to-day activities. • Mild AD At this stage, the cognitive signs become visible. Also, the pathological changes extend to the cerebral cortex. Symptoms include amnesia, forgetfulness, accompanied by inability to offer solutions to problems. Alterations in personality, mood swings and disorientations are also noticed. • Moderate AD Symptoms become more severe. The pathological alteration extends to the parts that control that speech, thinking and sense organs. Apart from the advanced severity of symptoms from the earlier stages, changes in behaviour and withdrawal syndromes begin to manifest. Those are accompanied by speech and imaginative impairment as well as inability to recognize family and friends. • Chronic AD At this stage, patients lose their control over daily activities. The pathological breakdown is said to spread over the cortex region. The symptoms include very poor cognitive abilities in patients, onset of systemic symptoms such as dyspraxia, malfunctioning of the sense of smell, trouble sleeping, akathisia, dystonia and parkinsonian symptoms. Further systemic symptoms start to appear, including difficulty performing learned motor tasks (dyspraxia), olfactory dysfunction, sleep disturbances, extrapyramidal motor signs like dystonia, akathisia, and parkinsonian symptoms. Alzheimer’s Disease Symptoms The symptoms of AD include the continuous decline in episodic memory and other cerebral skills, causing world cognitive deterioration, changes in character and conduct. There are early onset and late onset of Alzheimer’s disease. Late onset occurs in people above 65 years while early onset manifests in younger people. Dementia associated with AD in these subjects begins around 50 to 60 years. Maturity and poor educational background are the significant risk factors for late manifestation of Alzheimer’s disease. Others include the existence of the allele ε4 of the apolipoprotein E (APOE) gene, prolonged cardiovascular risk factors, past occurrence of brain shock accompanied by unconsciousness, inactivity, and poor cognitive demand. On the contrary, AD in younger subjects is commonly related to genetic mutations which are usually the Amyloid Precursor protein (APP) gene and Presenilin. Age doesn’t cause any disparity in clinical findings in AD patients. Akin clinical findings are recorded in AD patients irrespective of their age. The distinctive features of Alzheimer’s disease include the existence of neuritic plaques and neurofibrillary tangles, gliosis, neuronal loss and dystrophic neuritis as well. Neuritic plaques are exogenous lesions primarily consisting of -β42 peptide (Aβ42). Neurofibrillary tangles are endogenous wounds majorly made up of hyperphosphorylated TAU protein. Alzheimer’s Disease Treatments Current medications for Alzheimer’s disease are tacrine, galantamine, rivastigmine and donepezil (acetylcholinesterase inhibitors), NMDA receptor antagonist and memantine, though with little improvements in cognitive issues. The use of Aducanumad was recently permitted in the US. It is the first medical treatment approved for the disease because it works on the basic pathophysiology of the disorder. Acetylcholinesterase inhibitors act by decreasing ACh (Acetylcholine) catabolism which will increase the amount of acetylcholine in the brain and fight its loss due to cholinergic neurons. Mild to moderate side effects of these medications noticed in about 20% of its users include nausea and vomiting associated with cholinergic overload. The side effects of the medications can be controlled by regulating the dosage. Other rare side effects are bradycardia, low appetite, weight loss and more secretion of gastric acid. Similarly, EGb 761 (derived from Ginkgo biloba) is generally used for treatment of Alzheimer’s disease in Europe. It is active against symptoms of Alzheimer’s disease and vascular dementia. EGb 761 performs a vital function alone or in combination with other therapies to improve their efficiency. It protects the neurons, eliminates free radicals, boosts mitochondrial activities and regulates the level of serotonin and dopamine. Clinical studies for its effectiveness against mild and moderate dementia revealed significant reduction in cognitive problems, improved daily and mental activities. Nevertheless, EGb 761has not been proven to stop the development of Alzheimer’s disease to dementia. Alzheimer’s Disease: Amyloid Secretase Pathway Amyloid precursor protein (APP) is an integral polytopic protein that is among the major proteins in the central nervous system (CNS). It is also present in the epithelium and blood cells (peripheral tissues). On the other hand, amyloid precursor protein can be produced via two ways which are the amyloidogenic pathways and the non-amyloidogenic pathways. In the latter, the enzyme γ-secretase divides APP into sAPPα (soluble N-terminal fragment) and C83 (C-terminal fragment). The α-secretase acts on C83 to produce 3KDa (C3) (C-terminal fragment). Besides, β-secretase can divide Amyloid precursor protein into sAPPβ (smaller N-terminal fragment) and C99 (C-terminal fragment) that creates the full-length Aβ (β-amyloid peptides) due to further division by γ-secretase. There are different Aβ peptide species. Aβ40 and Aβ42 are majorly present in the brain. Aβ species are secreted as monomers which further become dimmers, trimers, protofibrils and fibrils which later accumulate, spread out and mature into amyloid plaques. Aβ40 and Aβ42 are closely related but Aβ42 is more susceptible to accumulation and fibrilization because it is most deadly Aβ peptide and performs a key function in the pathogenesis of Alzheimer’s disease. The Aβ oligomers are regarded as the most lethal form of the Aβ peptide. They relate with neurons and glial cells causing inflammatory cascaded, non-regulated calcium metabolism, oxidative stress, introduction of neuronal apoptosis and TAU phosphorylation. The combination of the aforementioned brings about a continuous beneficial response window where the release of Aβ peptides causes harmful effects to the neuronal cells. That adds to the malfunctioned APP metabolism and subsequent secretion of Aβ peptides. Amyloid-β fibrils accumulate in neuritic plaques chronologically listed: diffuse neuritic plaques, mature neuritic plaques, senile plaques and phantoms of senile plaques in severe Alzheimer’s disease. Also, the plaque build-up has a toxic effect on neuronal homeostasis causing neuronal malfunction and death. In the physiological circumstances, the APP is preferably broken down by the non-amyloidogenic pathway and a balance exists between the secretion of Aβ peptides and their removal from the cerebrum. Presently, Apolipoprotein E (APOE) and the insulin degrading enzyme (IDE) are the two proteins majorly responsible for the removal of Aβ peptides from the brain. The precise means by which Aβ peptides are removed from the brain is not explicit till date. However, a recent theory states that these proteins attach to Aβ peptides, preventing their accumulation into oligomers and ultimately leading to their removal from the brain. In the pathological circumstances, there is a metabolic change beneficial to the amyloidogenic division of APP which combines with reduced elimination of Aβ to cause the aggregation of Aβ in the brain tissue. Tau Protein TAU Protein Metabolism & Function The TAU protein is a microtubule related to protein existing in most tissues. It is majorly released in the central and peripheral nervous system. It is a key part of the neuronal cytoskeleton which enables it to work together with α- and β-globulin and make the microtubules become stable. In neurons, the microtubules are vital for the management of neuronal structure, synaptic plasticity, and axonal transport. Six TAU isoforms are present in humans. The key disparity among these isoforms is the differences in the N-terminal end of the protein and the presence of three or four domains attached to tubulin. The relationship between TAU and tubulin is a self-motivated process where TAU supports self polymerization and slows down rapid depolymerization of the tubulin. The procedure is controlled by the phosphorylation condition of TAU protein which is made up of about 79 phosphorylation spots at threonine and serine deposits. The equilibrium between phosphorylation and dephosphorylation of these epitopes supports conformational modifications that affect the way TAU protein relates with α- and β-tubulin and that calms the microtubules in the neurons. A lot of protein kinases and phosphatases are part of the control of TAU phosphorylation. Enzyme GSK3β (Glycogen synthase kinase 3β) is the central TAU-kinase in the neurons. A decreased action of certain phosphatases was also recorded in the cerebral tissues of subjects suffering from Alzheimer’s disease such as PP1, PP2A and PP5. The main deposits of serine and threonine that are present in a hyperphosphorylated state in harmful TAU and PHFs is accompanied by proline, indicating that TAU kinases is among the group of proline-controlled enzymes such as cyclin-dependent kinases (CDC2, CDK5, TPKll), MAPK (mitogen-activated kinases), and GSK3β. These kinds of enzymes can phosphorylate TAU outside a living organism (in vitro) and was found in the cellular strata of cerebral tissue of patients suffering from Alzheimer’s disease. Other kinases, such as PKC, casein-kinases I and II, calcium/calmodulin-stimulated protein kinase II (CaMPK-II) and protein kinase A (PKA), that are not directed by proline were also found in neurofibrillary tangles and are crucial to the control of TAU phosphorylation, adjusting the activity of the previous enzymes. At embryonic phase, neuronal TAU is primarily found in an agitated phosphorylated condition. This is because of the increased need for neuroplastic modifications in neurons and synapses at initial developmental phases of the central nervous system (CNS). However, in the full-fledged central nervous system, dephosphorylated TAU is needed to keep microtubules steady and manage structural and functional cytoskeletal homeostasis. Self-motivated alterations in the phosphorylation condition of TAU crops up in pathways used for neuronal changes and synaptic plasticity. For clinical diagnosis, TAU can be unusually hyperphosphorylated which affects its ability to attach to tubulin, causing the damage of the microtubule configuration, axonal movement and synaptic metabolism. Such alterations ultimately lead to cytoskeleton breakdown, failure of cell viability and neuronal death. TAU Protein & Alzheimer’s Disease NFT (neurofibrillary tangles) which are intra-neuronal wounds as described by Alois Alzheimer are majorly made up of paired helical filaments (PHFs) of hyperphosphorylated TAU. 25 unusual phosphorylation areas or more was explained in Alzheimer’s disease representing a sign of neuronal degeneration in this condition. The unusual phosphorylation of TAU in serine or threonine deposits close to the tubulin binding domain supports disaggregation of the TAU-tubulin complex and rearranges TAU into paired helical filaments to form neurofibrillary tangles (NFT). Owing to the significance of TAU in neuronal stability and homeostasis, its hyperphosphorylation brings about a flow of activities that eventually leads to neuronal death. GSK3β, Wnt Signaling and Alzheimer’s Disease In spite of reliable facts promoting the major role of Aβ peptides and hyperphosphorylated TAU in the pathogenesis of Alzheimer’s disease, none provide acceptable records for the complete range of pathological mechanisms related to this disease. NFTs and Aβ plaques may be a self-sufficient phenomenon in Alzheimer’s disease. So, other processes that were introduced that engage the action of upstream enzymes that are active in the aforementioned phenomenon. GSK3β is a vital enzyme in the maintenance of cellular metabolism including TAU phosphorylation. Wnt signaling is another essential arm of neurobiology of Alzheimer’s disease. Since Wnt signaling causes the deactivation of GSK, inhibiting TAU phosphorylation in the epitopes dependent on GSK. Regarding Alzheimer’s disease, GSK3β was discovered in a hyperactive condition causing the hyperphosphorylation of TAU. Besides, GSK3β controls APP metabolism and supports amyloidogenic division (for excess secretion of Aβ), decreases neurogenesis and amplifies apoptosis. AD and Other Related Diseases Alzheimer’s Disease and Dementia Alzheimer’s disease (AD) is the usual contributory factor to dementia among the aged. In the global aged population, about 24.3 million people experience dementia. In 2020, the statistics rose to 42.3 million people and expected to become 81.1 million by 2040. There are different facts that confirm the view that the malfunction in TAU homeostasis is a major occurrence in Alzheimer’s disease, an incident also present in similar neurodegenerative conditions like frontotemporal dementia and multiple system atrophy. Neuropathological research revealed that the circulation of NFTs in the brain correlates with the medical advancement of dementia in Alzheimer’s disease. Furthermore, intra-neuronal hyperphosphorylated TAU exists in the brain of patients with dementia that is not severe, without β-amyloid pathology. Hence, TAU hyperphosphorylation can be the initial occurrence in the physiopathology of Alzheimer’s disease (AD), while other processes including over secretion of Aβ, oxidative stress and start of inflammatory cascade may have lesser effects on the whole malfunction in neuronal homeostasis. Other Related Diseases Cardiovascular diseases such as diabetes, hypertension, and hypercholesterolemia are linked with increased risk of manifestation or rapid degeneration of Alzheimer’s disease. Other risk factors include smoking and hormone replacement therapy for menopausal women. • Berchtold NC, Cotman CW (1998). "Evolution in the conceptualization of dementia and Alzheimer's disease: Greco-Roman period to the 1960s". Neurobiology of Aging. 19 (3): 173–89. • Alzheimer A (1907). "Über eine eigenartige Erkrankung der Hirnrinde" [About a peculiar disease of the cerebral cortex]. Allgemeine Zeitschrift für Psychiatrie und Psychisch-Gerichtlich Medizin (in German). 64 (1–2): 146–48. • Berrios GE (1990). "Alzheimer's Disease: A Conceptual History". Int. J. Geriatr. Psychiatry. 5 (6): 355–65
Powered by Subscribe to the only car newsletter you’ll ever need Sustainable fuels: Formula 1 engines set to emit 65 percent less carbon by 2025 Zak Adkins The F1 championship wants a new sustainable power plant for 2025 that runs on eco-friendly fuel. In a bid to keep the emotive combustion engine alive in motorsport despite increasing concerns around carbon emissions, Formula 1 is intensively exploring lab-created fuels for a new F1 powerplant set to be deployed in 2025. The aim for the championship is to use a 100 percent sustainable “drop in” fuel that will run in a modern combustion engine with no major modification to existing powertrains.  The fuel will be created in the lab from non-food biomass, municipal waste or a carbon capture scheme.  Formula 1 ferrari All manufacturers, including Ferrari, will need to use the sustainable fuel from 2025. The plan is for Formula 1 to achieve greenhouse gas emission savings of up to 65 percent over fossil fuel alternatives once the sustainable fuel is deployed.  That figure is still some way off being 100 percent carbon free, but that will be difficult considering any combustion process must end with some form of tailpipe emission.  Obviously, electric motors provide a carbon free alternative, but Formula 1 does not seem willing to go down that path at this point in time.  F1 2022 car The 2022 Formula 1 car will use e10 fuel. Formula 1 says the new fuel will pack a punch, offering the same energy density as regular fuel. That means there won’t be any reductions in performance and race cars will be as fast as ever.  The key objectives of the 2025 power unit are to create environmental sustainability, implement a sustainable fuel, create an ‘emotive and powerful’ power unit, reduce costs and create an attractive engine spec for new manufacturers.  In the interim, F1 will move to E10 petrol fuels in 2022 ahead of a launch of the sustainable lab-created drop-in fuels later this decade. Formula 1 burnout Burnouts will still be possible, with Formula 1 saying the new fuel “packs a punch”.
How Technology Has Changed The Way People Play Sports James McArthur J. McArthur|09.10.16 Sponsored Links James McArthur September 10th, 2016 Over the recent years, technological advancements have proven to be of tremendous importance in almost every aspect of our lives. This includes education, business, healthcare, government and now the sports sector. Despite its impact in sports being not very significant in many parts of the world, it is evident that several aspects in the sports industry have changed and have in turn transformed the way people play sports. In this article therefore, I will discuss the various ways in which technology has changed the way people play sports. Technological improvements including application sensors, 3D modeling and motion capture has impacted greatly on the design of uniforms, gloves, helmets and pads to ensure comfort for the sportsmen and women. To be more specific, smart helmets have largely impacted on how players play. These helmets are specifically designed for players to protect them from the high risks of brain damage common in games such as rugby, hockey, skiing, and snowboarding among others. This helmet is specially made with a sensor and magnet technologies which serve to meet this purpose effectively. The sensors aid in the detection and dispersion of the force while the magnets function to displace as well as absorb the force. Therefore, where the player gets an injury while playing in the field, these helmets communicate to the coach in a unique way hence saving the situation. It does not just end there. Online bookmarkers like BonusCodeBets provide information about several aspects of how a team can perform before the match depending on how well the team has been working overall. Hawk-Eye technology This technology is useful in tracking ball movement. This technology utilizes quality video cameras (600 frames) located strategically during matches especially cricket, football, tennis and other ball games. Therefore even in cases where the ball has been blocked from the cameras, Hawk-Eye technology is able to trace the ball thereby facilitating efficient and effective decision making by the referees. In addition, this technology is useful to players as it provides ground for their training on a player- by- player basis hence helping in the improvement of individuals' weak points. Tools of training Sports training have revolutionized to include videos as well as technologies in body measurements. In addition, computer software have also been developed in this field and have proven to be useful in helping sportsmen connect with their body workings the best ways possible while also ensuring fitness and proper nutrition. Such software include: BodyByte: This computer software is useful in organizing and managing all information associated with the players' nutrition, fitness and training. Teem Beep Test: This computer software is useful I performing and recording beep tests into your personal computer. In conclusion, it is important to note that with many people venturing into sports as a career, technology now has a major space to explore and make life easier for the players, referees as well as the crowd. This entails coming up with technologies that not only meets the need of a single individual but all the parties involved. This will not only lead to high quality games but also fair judgments by the parties involved. Therefore with the above technological changes in place within the sports industry, it is evident that technology has for sure changed the way people play sports. Popular on Engadget
More From Forbes Edit Story Editors' Pick| Kelly Phillips Erb Sin taxes are excise taxes imposed on goods or behaviors - like booze and cigarettes - that lawmakers deem harmful. In addition to raising revenue, the idea is that bumping taxes high enough should trigger a slowdown in the behavior. But what happens if taxpayers simply exchange the "sinful" behavior - not for a "better" response - but for another bad behavior? That's precisely what a new study funded by the National Institutes of Health suggests: raising taxes on e-cigarettes in an attempt to curb vaping may cause people to purchase more traditional cigarettes. Economists talk about the response to sin taxes in terms of elasticity. If the idea is to reduce bad behavior, it's critical to figure out how high you can raise the tax to get people to reduce consumption. For example, when studying soda taxes, economist Roland Sturm concluded that "Small taxes will not prevent obesity." Sturm determined that taxes would need to hit about 18 cents per dollar to affect behavior. The same general idea was used to look at e-cigarettes. A team of researchers from six universities examined the effect of e-cigarette taxes enacted in eight states and two large counties on e-cigarette prices, e-cigarette sales, and sales of other tobacco products. Using data from 35,000 national retailers from 2011 to 2017, researchers found that for every 10% increase in e-cigarette prices, e-cigarette sales dropped 26%. But the same 10% increase in e-cigarette prices caused traditional cigarette sales to jump by 11%. And while that sounds simple, it's not that easy. First, these aren't exactly apple-to-apple comparisons. E-cigarette products can vary wildly: some are disposable e-cigarettes, while others are starter kits or refill cartridges. And, the study notes, the number of cartridges, liquid, and nicotine may be very different - even within products of the same type. The corresponding taxes are just as mixed with some levied in proportion to the liquid volume in each e-cigarette product, while others are ad valorem taxes (meaning based on the value). In contrast, traditional cigarette taxes are typically levied in terms of dollars per cigarette. The math is even more complicated. It's the stuff of data analysis flashbacks/nightmares: But the conclusion, according to the study, may be simple: e-cigarettes are an elastic good. And since e-cigarettes and traditional cigarettes are economic substitutes, e-cigarette sales may increase as traditional cigarette taxes go up, and conventional cigarette sales may increase as e-cigarette taxes rise. Currently, nearly half of the fifty states have an e-cigarette tax, which has significantly raised the price of e-cigarettes. Congress is considering enacting a federal tax of its own on e-cigarettes. That concerns study co-author Michael Pesko, an economist from Georgia State University. "We estimate that for every e-cigarette pod no longer purchased as a result of an e-cigarette tax, 6.2 extra packs of cigarettes are purchased instead," he said. "Although vaping-related illnesses are a public health concern, cigarettes continue to kill nearly 480,000 Americans each year, and several reviews support the conclusion that e-cigarettes contain fewer toxicants and are safer for non-pregnant adults," said Erik Nesson, an economics professor in the Miller College of Business at Ball State, and a member of the research team. According to the Centers for Disease Control and Prevention (CDC), nearly 3% of adults in the United States used e-cigarettes in 2017. Using e-cigarettes - or vaping - among adolescents has grown even more rapidly, with almost 27.5% of high school students using e-cigarettes in 2019.  And while vaping-related illnesses are a public health concern, according to the CDC, traditional cigarettes continue to kill nearly 480,000 Americans each year. Some studies suggest that e-cigarettes contain fewer toxicants and are safer for non-pregnant adults than conventional cigarettes.  That poses an interesting dilemma. If raising taxes on e-cigarettes - the long-term effects of which are not wholly known - simply chases smokers to traditional cigarettes - which the CDC has determined to be "the leading cause of preventable death" - is it worth it? The research team believes that the issue will continue to be important for policymakers to consider as they develop e-cigarette related tobacco control policies. You can read the study, "The Effects of E-Cigarette Taxes on E-Cigarette Prices and Consumption: Evidence from Retail Panel Data," from the National Bureau of Economic Research. The research team was made up of six economists: Dr. Erik Nesson at Ball State University, Dr. Nathan Tefft at Bates College, Dr. Michael Pesko at Georgia State University, Dr. Catherine Maclean at Temple University, Dr. Charles Courtemanche at the University of Kentucky, and Dr. Chad Cotti at the University of Wisconsin Oshkosh. Follow me on Twitter or LinkedInCheck out my website
Tips for Improving Your Dental Health Tips for Improving Your Dental Health Tips for Improving Your Dental Health Bad teeth are harmful to your well-being. The relationship between good dental hygiene and your overall health has been proven to go hand-in-hand. Therefore, it is imperative that you develop habits to maintain a healthy smile. Here are some steps that help you improve your dental health. Brush Teeth Regularly Be sure to take two minutes to brush your teeth at lease twice a day. Your brushing technique is also important. An up and down motion is the best way to brush your teeth, because the brushes will be traveling across the natural slopes of your mouth. Do not scrub vigorously, rather brush thoroughly and gently. Scrubbing your teeth too hard can damage the tooth enamel and scratch your teeth. The toothpaste you use should contain fluoride in order to prevent cavities. Some toothpaste contains whitening agents, which can help your teeth become brighter, whiter and more attractive. However, whitening toothpaste can also contain abrasives, so a gentle yet meticulous brushing technique should be used. Purchase the Correct Toothbrush One of the first things you can do to improve your dental health is purchase the correct toothbrush. Soft toothbrushes are easy on the teeth; yet still provide a thorough cleaning. Electric toothbrushes also provide a deep cleaning while preserving the enamel of your teeth. Stay away from firm toothbrushes, as they tend to scratch and scrape the teeth and gums. Ask your dentist what he or she recommends for your specific needs. Flossing is also very important for dental health, and many of us have had the lecture from our dentist about not flossing often enough. Try to set aside a minute or two to floss your teeth before bed each night. Flossing removes food particles that can’t be reached with normal brushing. Without removal, these particles will begin to decay, causing bad breath and gingivitis. An antiseptic mouthwash can also be used on a daily basis to kill bacteria and freshen breath. Chewing a sugar-free gum after a meal is another way to remove loose food particles from teeth, preventing staining, halitosis and cavities. Can teeth heal from cavities? Cavities, also called dental caries, are the result of untreated tooth decay. Over time, this tooth decay can cause holes that go deep into your teeth, sometimes all the way to the root. Luckily, early cavities can be prevented — and even heal on their own — with the right approach to oral hygiene. How do you take care of your teeth when you have cavities? * Ask your dentist which tips are best for you. * Brush with fluoride toothpaste after eating or drinking. * Rinse your mouth. * Visit your dentist regularly. * Consider dental sealants. * Drink some tap water. * Avoid frequent snacking and sipping. * Eat tooth-healthy foods. * Consider fluoride treatments. How do you heal a cavity in 2 days? If you have a cavity, there’s only one way to get rid of it in two days: seeing your dentist right away. To repair cavities, dentists perform a simple filling procedure. First, the dentist removes the decayed portion of the tooth. Then, they clean the area and place a filling material where the cavity was in the tooth. How can I heal a cavity naturally? * These remedies can help to provide some relief and support your oral health. * Oil of Oregano to Stop Decay. * Establish a Healthy pH Level with Xylitol. * Clove Oil for Pain Relief and Prevention. * Neem Bark to Fight Infection. * Green Tea as a Natural Fluoride Treatment. * Keep Teeth Healthy with Licorice Root. What foods cause cavities? * Chewy Candy. Eating lots of sugar is bad for oral health, and candies are full of it, making them one of the most common foods that are bad for your teeth. * Starchy Snacks. * Citrus Fruits and Sugary Juices. * Sweet Treats.
Javatpoint Logo Javatpoint Logo Artificial Intelligence in Cybersecurity: Pitting Algorithms vs Algorithms Artificial Intelligence in cybersecurity is revolutionizing how we look at protecting ourselves against dark-hat hackers and hackers with malicious motives. It's important to take a focused strategy when it comes to cybersecurity. From a technical standpoint, it's important to take a comprehensive approach to security, where companies know the various algorithms available to create an enhanced security ecosystem. Even the largest security team comprised of the most skilled IT professionals can't sift through the thousands of vulnerabilities to determine which to prioritise. Tools leveraging Artificial Intelligence that continuously monitor all assets, proactively predict what vulnerabilities are most likely to be exploited and produce a prioritised list of fixes are entirely necessary to keep up with constantly evolving attack methods. - Gaurav Banga, CEO and founder of Balbix It is essential to know the best way to implement anti-attack measures to ensure that every attack can be identified and stopped at a rapid time. Since timing the actions is essential, companies must expand their cyber security procedures to ensure there aren't any lapses in the steps implemented. This is the case for smaller to mid-sized companies focused specifically on security from a technical perspective. They do not create a value-based system that can be an important resource for the company. They rather focus on constructing strategies that focus on engaging on a necessity-only basis. It's the reason algorithms are being created to enhance the quality of security. In a time when cybersecurity is a must, it's crucial to take an increased position. Algorithms are the sole tool that will allow businesses to improve their cybersecurity measures. In addition, businesses can improve their current security by leveraging third-party companies that can create excellent strategies for you. This makes the process shorter and helps in better control of cybersecurity threats. The Evolving Nature of Cybercrime The field of cybersecurity, it's growing rapidly. It is due to attacks becoming more sophisticated and more attackers emerging out of the apex. This is why cybersecurity professionals have to be aware of how cybersecurity evolves. Cybercrime is shifting to newer trends, such as crypto-jacking and ransomware, at increasing alarming rates of growth. Cybercrimes have been developed to capture information from various sources and publish data through dark websites. Many of the most renowned tech firms of all time have been compromised, and countermeasures are implemented every day. But, there must be a more comprehensive approach to cybersecurity. That is the kind of thing that Artificial Intelligence in cybersecurity can provide. While hackers employ their algorithms to attack systems, cybersecurity professionals employ their own. This is the reason the battle of algorithms is waged, with technology being employed to its fullest potential. The stereotype of young guys in hoodies has been cast out for some time, but under-representation means the sector is still portrayed as a job for the boys. As just one example only 11% of the cybersecurity workforce are women. But are set to see the needle shift. From the language that determines what we consider to be 'cybersecurity' through to the numbers entering the profession, we will start to see more of ourselves engaged in, and reflected in, the cause to keep the digital realm trusted and safe. - Lydia Ragoonanan, The Director of LORCA Artificial Intelligence is also enabling technologists to create an internet that is more secure. A world where information is safe and leaks aren't a problem can be a positive factor for the good of the business. Societies are also safer as they know that hackers don't have access to the systems that hold sensitive data. Because information is a valuable resource, companies must use top-of-the-line security to make sure there aren't any leaks or data leaks. Better Prediction through Artificial Intelligence This is an excellent tool that lets us discover vulnerabilities that might be in the cybersecurity realm. It aids us in making more accurate predictions in relation to crime. Additionally, it allows cyber-security experts to determine their next steps before the attack. It is also possible to have attacks on day 0, which are passive. They can be stopped and eliminated from the system through running Artificial Intelligence algorithms. Because the number of victims is huge and there are many possible endpoints, it's suggested to employ the AI algorithm in order to guarantee adequate security. AI model. Predictive analytics can also assist us in gaining a legal perspective. It lets cybersecurity experts view hackers from a different view. It provides a deeper understanding of security that is platform-centric to cyberspace. It also lets us discover new ways of protecting ourselves based on the predictions that are made by AI models. Artificial Intelligence, in turn, will also be able to create better systems that actively look for weaknesses. Hackers will be stopped from entering since more and more entrances are shut off. They also have the ability to maintain an ongoing supply of AI-enabled devices that will continue to play the role of security threat detection. The overall model gets better as time passes because the AI system gains knowledge from the information it is presented. The platform is taking a more proactive approach to enable AI systems to extend their full potential. In addition, there is a greater emphasis being placed on the AI systems that are part of the platform. This provides them with more control over their systems and allows for an automatic response following detection. Also, the AI system is able to detect and take action to ward off threats without the need for intervention. As Artificial Intelligence software gets more advanced, the capabilities of AI expand as well. Enhancing the Scope of Artificial Intelligence in Cybersecurity Companies must investigate the potential of AI for their cybersecurity infrastructure. AI can improve cybersecurity while also putting a greater concentration on communication. Research suggests the majority of security alerts go without being noticed, while less than 23% of organizations respond only in the time of a crisis. The concept of cybersecurity must change to provide a complete strategy for security management. It also implies that there must be a greater emphasis on AI for cybersecurity. With the introduction of new technologies and advanced algorithms, there will be a race to see who is more powerful. When AI is able to go through all the files of a system and spot dangers before they appear in the system, the entire ecosystem gets substantially more robust. Furthermore, there are more opportunities for teams in general since AI provides more details in the form of feedback. More tools for cyber-teams can incorporate the data back to the AI algorithm. This improves the performance of the platform can be used the possibility of greater Machine Learning integration. At present, approximately 20% of executives in the C-suite use ML to improve their AI offerings. This is why there must be an expansion of the range when it comes down to AI and its capabilities. AI must have a greater range and more coordinated approach to threat detection. To realise the full potential of AI, we must integrate its various forms to offset the limitations of each. No one approach will be sufficient because each approach is optimised for one specific set of problems at the expense of others. - Scott Lathrop, Senior Researcher AI. In terms of the capabilities of Artificial Intelligence, companies must examine their current technology. They must assess whether they are prepared with regard to Artificial Intelligence and their cybersecurity measures. In addition, there is a requirement for cybersecurity experts to increase the reach of each business. The business needs to be more active regarding cybersecurity strategies. Also, it must add greater control over the elements which are not related to cybersecurity. Artificial Intelligence in Cyber Policies To ensure the correct execution of policy and Artificial Intelligence enhanced safeguards, AI could be employed to create better rules. AI can be utilized to increase the work capabilities that are carried out by the AI platform and to enforce more strict rules at the same time. When employees commit a lapse or upload a file, they aren't supposed to be sharing. The AI system will automatically recognize it as an anomaly and block them from doing it. This is one of the less-discussed functions in the realm of cyber-compliance. AI can be utilized to help reinforce and set the guidelines by which businesses can increase their employees' compliance. By implementing specific procedures to ensure compliance locally, AI can enable companies to be more secure from the perspective of non-technological methods. It also allows businesses to increase the worth of employees by making their workplaces more trustworthy. The security net provided by AI and cyber-compliance permits employees to explore more and develop creative solutions. This is a great way to aid the development of business without noticing any obstacles from long processes and procedures. It will automatically recognize the moment a mistake was committed and attempt to repair it independently. Cyber governance is a new area within the AI and cybersecurity arena and focuses on covering the greater terrain. All employees and end-points are active and more aware of the possible problems in this domain. This can also be an important way for employees to be aware that they've committed an error and try to correct the issue at a faster speed. Artificial Intelligence can assist in the formulation of these cyber-policies as well as enforcing their rules from a technological perspective. It takes human errors out of the equation and creates a stronger platform for companies to analyse. Artificial Intelligence dashboards can be utilized to make sure that there aren't any technological gaps. There are other areas where AI could be able to enforce greater control over particular models. This is where Artificial Intelligence could become the dominant force and come up with creative solutions. Strengthening Models of Defence The primary need is to enhance the impact of cases we currently have within the corporate sector. We do not have the advanced tools required to build better defence mechanisms for the fundamental products. From consumer information to long procedures, the cost of cyber security increases each day. Businesses must adhere to the legislation and rules that regulate handling more information. This is where the concept of making use of AI to combat hackers came from. Hackers are using advanced technologies; however, AI has ensured that the battle remains calm. This is because AI is constantly being utilized to enhance all business areas. If companies are better prepared to face the world at a higher level and are more prepared, they can fight in all areas. Most of the time, it only takes one outdated computer or an insecure network for hackers to gain access. This is why this technology is so beneficial. Technology can be implemented and revived without requiring many changes, especially with cybersecurity technologies being developed every year. The strength of core models is also dependent on improving the technology itself. Technology has to become more sophisticated and sophisticated to keep pace with evolving requirements. AI is a process that can be repeated over, but a clear solutions model has to be created to ensure that the system is up-to-date. There are a variety of experts and third-party integrations which allow for greater coverage within the AI field. However, it is essential to develop an organized strategy for AI within cyber. In addition, businesses need to recognize the value AI brings into the equation. This is when they begin to build on existing resources to provide a more secure environment. AI within cyberspace is vital to be executed correctly since the scope must align with the goals of the business. AI should not be viewed as an easy-to-use model. It should be a long-term objective that businesses must achieve. Clouds, VPN, and AI Cybersecurity As cloud computing has grown in adoption over the past few years, there are new problems that are emerging within the AI cybersecurity field. Many companies are launching innovative tools that offer better solutions to facilitate communication. Cloud is a single platform that has altered the way we think about technology and networking, as well as AI, being the next to follow. This has resulted in an increase in the use of other technologies, such as VPNs, as well as Machine Learning. AI is a new field that needs to be just one step behind hackers to defend new technology and platform. From a cloud standpoint, the inter-connectivity level is amazing. The number of devices connected to the cloud now than before makes the task of cybersecurity much more challenging. This has resulted in an unavoidable increase in the use of AI and has made cybersecurity more possible across the cloud. Clouds are also more secure in the event that AI provides a security layer for all transactions and servers. In conclusion, you can get more security and more protection for vendors and third parties. VPNs are also intended to offer security when it comes to privacy and security. However, as the growth of VPN continues, there's more requirement for cybersecurity too. As more people use VPNs are at a greater risk of being a target for hackers to make use of. This could lead to new problems arising while relying heavily on VPNs to ensure that the VPN tunnel to be secured and safe. AI to investigate hacking behaviours An even more thorough method is to implement AI throughout the entire process. Businesses can be relaxed because AI technology provides greater security. It can also connect to networks across the internet and ensure they're secured. The dashboard's core features give all the data needed, which eventually leads to a more cohesive platform. Cyber compliance officers can manage the field with growing confidence in the capabilities of the technology. AI cybersecurity is also a unique perspective to the area. Many companies are developing new algorithms. Only a few of them are analysing the hacker's attempts. They're making use of AI as a means of research to find out the ways and reasons that pirates are able to escape. They are also able to employ AI to develop more effective traps and test mechanisms for hackers to fall for the trap. They can also study the scam from a technical perspective and discover new information they might have missed. AI experts will then be able to design more secure security measures after analysing how hackers were able to get away using the information they had stored. Utilizing AI in the realm of research assists in strengthening certain areas too. Audits can be conducted by companies to ensure that their company is safe from threats and then develop solutions to keep them safe. This is a good method of ensuring that all necessary measures are in place, particularly in compliance with all laws. Cybersecurity is strengthened by investing in AI research. AI for cybersecurity is a new idea that's being explored by algorithmic. Businesses are pushing to develop a complete tool to protect themselves from cybercriminals. From trojans to spyware, there is a need to be more proactive in monitoring and protection. If not, companies will be at a disadvantage in defending against such attacks. There is also enormous scope for the development of AI to include new areas. By utilizing Machine Learning and Big Data, AI is opening up boundaries to expand the scope for what cybersecurity professionals can accomplish with their resources. From technology to talent, there are a variety of areas in which AI can help cybersecurity professionals increase their capabilities. Youtube For Videos Join Our Youtube Channel: Join Now Help Others, Please Share facebook twitter pinterest Learn Latest Tutorials Trending Technologies B.Tech / MCA
In quantum mechanics, not even time flows as you might expect it to University of Vienna Forward and backward time flows When looking at the celestial motion, one often perceives a sense of eternity that could lead us to wonder whether time really exists after all. On the other hand, looking back at our daily lives extinguishes all doubts: time exists and moves forward. This apparent certainty comes from the fact that most macroscopic physical phenomena can only ever occur in one direction. Take the sequence of things we do in our morning routine, for example. If one were to show us our toothpaste moving from the toothbrush back into its tube, we would know without a doubt that they were just showing us the rewind of the recording of our day. In physics, this propensity of certain phenomena to occur in only one time’s direction is linked to their production of ‘entropy’, which is the physical quantity defining the amount of disorder in a system. In nature, processes tend to evolve spontaneously from states with less disorder to states with more disorder, and this propensity can be used to identify an arrow of time. Thus, if a phenomenon produces a large amount of entropy, observing its time-reversal is so improbable as to become essentially impossible. However, when the entropy produced is small enough, there is a non-negligible probability of seeing the time-reversal of a phenomenon occur naturally. Thinking back to the toothpaste example, if we were to squeeze the tube only gently and only a very small part of the toothpaste came out, it would not be so unlikely to observe it re-entering the tube, sucked in by the tube’s decompression. On the other hand, as the tube is squeezed with more strength, the toothpaste will spread out in an irreversible way, requiring a much greater effort if one were to put all it back in. The boundary between forward and backward blurs in quantum mechanics A team of physicists at the Universities of Vienna, Bristol, the Balearic Islands and IQOQI-Vienna, under the lead of Časlav Brukner, has applied this idea to the quantum realm, seeking to gain a deeper understanding of how time flows in that regime. One of the peculiarities of the quantum world is the principle of quantum superposition, according to which if two states of a quantum system are both possible, then that system can also be in both states at the same time. If we look back at the system evolving in one or the other temporal direction (the toothpaste coming out of or going back into the tube), it results that quantum systems can also find themselves evolving simultaneously along both temporal directions. Although this idea seems rather nonsensical when applied to our day-to-day experience, at its most fundamental level, the laws of the universe are based on quantum-mechanical principles. This begs the question of why we never encounter these superpositions of time flows in nature. “In our work, we quantified the entropy produced by a system evolving in quantum superposition of processes with opposite time arrows. We found that this most often results in projecting the system onto a well-defined time’s direction, corresponding to the most likely process of the two” explains Gonzalo Manzano, a co-author of the study. And yet, when small amounts of entropy are involved (for instance, when there is so little toothpaste spilled that one could see it being reabsorbed into the tube), then one can physically observe the consequences of the system having evolved along the forward and backward temporal directions at the same time. As pointed out by Giulia Rubino, lead-author of the publication, “although time is often treated as a continuously increasing parameter, our study shows that the laws governing its flow in quantum mechanical contexts are much more complex. This may suggest that we need to rethink the way we represent this quantity in all those contexts where quantum laws play a crucial role.” Publication in Communications Physics: Quantum superposition of thermodynamic evolutions with opposing time’s arrows, G. Rubino, G. Manzano and C. Brukner. Communications Physics (2021). DOI: 10.1038/s42005-021-00759-1
Parotidectomy or submandibular gland excision The parotid gland is the largest of the salivary glands and lies on the cheek, just below the ear. The commonest reason for parotid surgery is the presence of a lump within the gland. The majority of these are benign, but they require removal to conform the diagnosis and to remove the risk of future cancerous change. Surgery is performed under a general anaesthetic, and involves removal of some (partial parotidectomy) or all (total parotidectomy) of the gland. The affected area of the parotid gland is removed preserving critical structures in the area including the facial nerve. Further information on parotid surgery and parotidectomy from ENT UK is available here. The submandibular glands lie just below the jaw. Surgery to remove the submandibular gland is usually performed for recurrent infection, stone disease or tumours. It involves a general anaesthetic and an incision in the skin of the neck. The whole submandibular gland is removed, preserving important nearby nerves and blood vessels. Further information on submandibular gland surgery from ENT UK is available here. After the operation a small drain is inserted, which is removed the next day. In most cases both parotidectomy and submandibular gland excision require an overnight stay in hospital. Skin incisions are usually closed with dissolvable stitches, which do not need to be removed. Patients are seen in clinic 2-3 weeks after surgery for a check up and to explain the laboratory results from any surgical specimens.
The effects of daylight on chemistry Discussion in 'Darkroom Developing and Printing' started by John, Oct 12, 2003. 1. John John Guest Alrighty, just came across some fixer that was displaced during the move to our new home (see ) It was stored in a 5 gallon tank with a floating lid. Unfortunately it's also been stored outside in daylight since June. The tank is about half full. Anyone care to explain the reactions that happen when a chemical such as thiosulfate is exposed to high levels of light for that length of time ? FWIW, I haven't tried it yet as I've been working a lot lately. Test strips tonight. John, Oct 12, 2003 1. Advertisements 2. John Gary Beasley Guest As far as I know thiosulfate will not be affected by the sunlight. Thiosulfate is used in pool maintenance to help regulate the chlorine levels so it it regularly used in brightly lighted conditions. Heat and/or ph problems most likely will be the culprits to cause sulfur to precipitate out. If you see no sulfur you are probably fine. Developers and silver salted solutions have always been recommended to be kept out of bright light, the first because organic compounds can be broken by energetic photons and silver salts go without saying they are sensitive. Gary Beasley, Oct 12, 2003 1. Advertisements 3. John David Foy Guest David Foy, Oct 14, 2003 1. Advertisements Ask a Question Want to reply to this thread or ask your own question?
Alex N. Gecan Asbury Park Press ASBURY PARK - Jersey Shore residents of all ages gathered together Saturday for the nation's first official Juneteenth celebration, commemorating the end of slavery in Confederate states 156 years ago. Organizers in Neptune and Asbury Park had combined efforts across Ridge Avenue for a joint weekend-long Asbury Park & Neptune Juneteenth Celebration. Local vendors hocked their wares and community groups handed out information throughout a mostly bright, sweltering Saturday at Springwood Park in Asbury Park. "A lot of people didn't know about it," said Sharon Simmons, who had come to Asbury Park from Tinton Falls for the festival. "We never heard about it in school, never." Simmons said her church, Red Bank Calvary Baptist Church, has put on Juneteenth skits for past celebrations, and that she had gone there Saturday morning. "Now we know what it means. We're grateful for the holiday," Simmons said. The holiday dates back to June 1865, months after the Civil War had ended. Enforcement of President Abraham Lincoln's Emancipation Proclamation nearly three years earlier had not reached all corners of the former Confederate States of America. U.S. Maj. Gen. Gordon Granger finally brought news of freedom to slaves around Galveston, Texas, in June 1865. Slavery persisted in the United States after that, however. Delaware and Kentucky continued to allow slavery for nearly six months more. The Emancipation Proclamation had applied only to Confederate states, not to border states which stayed in the Union. Legislatures in both Delaware and Kentucky rejected the 13th Amendment, which formally abolished slavery, after Congress passed it in 1865, but the amendment was nevertheless ratified that December. Jersey Shore residents gathered at Springwood Park in Asbury Park June 19, 2021, to commemorate Juneteenth. "It's been a long time coming, and it's about time," said Diane Wran-Farris, a Neptune resident. "We're so grateful, because the celebration of this brings people to mind of what our people have been through the very beginning." Wran-Farris pointed to government buildings in Washington, D.C., that were the products of slave labor as proof that Africans, African-Americans and other Black Americans had literally built the country that had enslaved them. "As the offspring of these individuals who sacrificed so much so that we could have freedom, so that we could have equality, and the challenge right now that we're up against is maintaining that and getting it back, because there are factions that want to take it away from us. They're afraid of our growth," Wran-Farris said. President Joe Biden signed a law Thursday making Juneteenth a national holiday. But in New Jersey, Gov. Phil Murphy declared Juneteenth a state holiday in September 2020. "We never really had awareness of it, but now the children of this generation will grow up being used to Juneteenth, just like Columbus Day or the other holidays that are out here," said Kelli Sanders, an Asbury Park native who now lives in Brick. Asbury Park itself was founded by segregationist James A. Bradley. "I believe there is still a way to go. I'm born and raised in Asbury Park and there was always a separation by the tracks," said Sanders, referring to the train tracks that run parallel to Memorial Drive. "The people of color lived on one side of the tracks, and people — Caucasian and so forth — lived on the other side closer to the beach." "It's a beautiful thing, and we need to do more of this," said Sanders' mother, Rovena Sanders, who lives in Asbury Park and had come out Saturday with her daughter. "Bring everybody together. Everybody needs to be together."
How Exercise Reduces Cholesterol Levels, Heart Attacks, And Strokes | Mukaila Kareem, DPT | RxEconsult All Health Articles How Exercise Improves Cholesterol Levels And Heart Disease Risk Category: Cholesterol by - October 26, 2015 | Views: 41755 | Likes: 0 | Comment: 7   Cholesterol Transport Cholesterol is a lipid and, therefore, does not dissolve in water. Therefore, it needs a carrier or “shield” for successful transportation in the water-soluble blood environment. This is achieved by specialized transport particles called lipoproteins. There are 6 types of lipoproteins but for the purpose of this article, low-density lipoprotein (LDL) and high-density lipoprotein (HDL) are the specialized particles responsible for cholesterol transport that will be discussed.  The liver is the powerhouse in cholesterol metabolism and is largely responsible for its production, storage and the only organ for its elimination. In a two-way traffic, LDL particles deliver cholesterol to all parts of the body where it is needed while HDL particles remove excess cholesterol from peripheral tissues back to the liver for excretion or recycling. The process of cholesterol removal from other tissues back to the liver is called reverse cholesterol transport. Clinical Implications Of High Cholesterol Levels Plasma LDL-cholesterol can rise significantly with age, diet, and as well as genetic defects. A study indicated that plasma LDL-cholesterol can increase by about 40% between 20 and 60 years of age. Adult western diets contain between 30% and 40% of fats and the substantial amount of saturated fats in these diets may contribute to increased plasma LDL-cholesterol. With high plasma LDL-cholesterol concentrations, some LDL-cholesterol will attract and bind with immune cells in the arterial wall causing plaque and narrowing of blood vessels. For this singular reason, LDL-cholesterol is notoriously called the “bad cholesterol”. In fact, the clogging of coronary arteries supplying the heart muscles causes heart attacks and has caused more death and disability than all types of cancers combined! This does not include the incidence of ischemic strokes (brain attacks) that occur from reduced blood flow to the brain primarily related to “bad cholesterol”. According to the CDC, more than 73 million adults have high LDL-cholesterol and only about 30% have it under control. Every year, about 610,000 people die of heart disease in the United States. Next: Effect Of Exercise On Cholesterol Levels And Heart Disease Risk For More Healthcare Insights Join Us On Twitter and Facebook. Join The Community To Publish Articles.
Menopause is not a disease in itself, but it is accompanied by bothersome and embarrassing symptoms for many women. Menopause is triggered by the significant decrease in the body’s natural production of estrogen. The main symptom remains hot flashes, a sudden sensation of heat initiating from the chest and going up towards the face. When hot flashes happen at night, they are called night sweats. Irritability, mood swings and vaginal dryness are other frequent symptoms of this transition period. Many symptoms of menopause can be controlled with the proper lifestyle. Between adequate hydration, regular physical exercise and stress management, diet remains a major impacting factor on the incidence of menopausal symptoms. Some foods can actually increase the number of hot flashes while other foods will facilitate the transition to post-menopause. Among foods that can worsen the situation, we find strong spices, alcohol, sugar and saturated fats. On the other hand, phytoestrogen-rich foods such as soy, flax and legumes will help rebalance the hormonal activity in the female body. It is well known that Asian women experience fewer, less severe menopause symptoms than Western women, in part due to their lifestyle and phytoestrogen-rich diet, especially from soy. Eating smaller meals will also be beneficial. Simply add healthy snacks between them to get your fill.  A menopausal woman should drink between 1.5 and 2 liters of water daily. Not only will a good hydration prevent tissue dehydration (which is more frequent at this time of life), but it will also minimize the impact of hot flashes.  A menopausal woman should remain physically active. The general health will be improved, but some sources believe exercise can reduce the incidence of hot flashes. Smoking and physical inactivity can increase the frequency of hot flashes. It is therefore important to remain active even during this time of life and it is always time to get rid of bad habits such as smoking.
Root Rot in Monstera: How to Diagnose and Save it Monstera is a beautiful plant to keep in your home. Unfortunately, like many plants, it is susceptible to root rot if it is not being given ideal conditions. Unless it is caught, root rot will end up killing your beloved Monstera.  Fortunately, this comprehensive guide to root rot in Monsteras will help you keep your plant healthy. In this article, we have everything you could possibly want to know about root rot in Monsteras: Ready to learn how to keep your Monstera safe and happy? Let’s get started!  Disclaimer: As an Amazon Associate I earn from qualifying purchases. What is Root Rot? This is how a perfectly healthy Monstera’s root should look like after rinsing with water. Root rot is a fungal infection in the soil caused by overwatering. Too much moisture in the soil for too long creates an anaerobic environment, which will then cause your Monstera’s roots to begin to decay and rot.  Once root rot takes hold, it starts by destroying the plant’s roots and works its way up to the main stem.  The roots turn from white to black and slimy, and eventually die. If not taken care of, root rot will kill a Monstera quickly.  As the roots are under the soil, it is hard to notice root rot unless you pay attention to other signs your Monstera gives you that it is not healthy.  Root rot can be caused by various factors, all of which have to do with there being too much water in the soil for too long. Plants need air just like we do! Root rot can be caused by chronic overwatering or if the soil has already been contaminated by fungal root rot, a single occasion of overwatering.  What Causes Root Rot in Monstera? So now we know what root rot is, let’s explore what causes it.  We know that root rot loves anaerobic soil caused by excess moisture. Unfortunately, there are lots of ways for this to happen and for infection to take hold.  Overwatering is the main culprit and the most common way for a Monstera to fall victim to root rot.  Overwatering can be caused by a few different things, all of which cause more water to accumulate in the soil than your Monstera can drink.  Your Monstera’s soil should be moist like a wrung-out sponge, not soggy or wet to the touch.  Some causes of overwatering: • Watering too frequently • Not enough drainage in the pot • Too much clay or organic matter in the soil • Pot is too big Watering your Monstera too frequently can cause it to become overwatered and develop root rot.  To prevent yourself from watering your Monstera too often, you need to check its soil moisture and only water it once the top two inches of soil are dry.  When you water your Monstera, give it lots of water until it starts running out the drainage hole in the bottom.  The amount of water you give your Monstera in one sitting will not affect its likelihood of developing root rot, as long as the soil has dried out enough. Overwatering only occurs when you watered your Monstera when the soil was already wet.  Also read: Complete Guide to Watering a Monstera – How Often Should You Water? Not enough drainage in your Monstera’s pot can also contribute to overwatering.  A pot without proper drainage in the bottom will cause water to collect in the bottom of the pot. If this water sits too long without being absorbed by your Monstera, the soil and roots will start to rot in the anaerobic environment.  Too much clay or organic matter in the soil will prevent water from draining as well. Too much water sticking around for too long can cause root rot. You want your Monstera’s soil to drain freely.  Garden soil is not suitable for potted plants like Monsteras because it retains too much moisture and can contain pathogens that cause root rot and other diseases or pests.  A pot that is too big can also be a cause of root rot. When your Monstera’s pot is too big, there are soil sections that contain no roots to take up water.  Those root-less sections of soil will stay wet for a very long time until the water naturally evaporates. If it takes too long to evaporate, the soil may begin to decompose and cause the roots to start rotting as well.  When repotting your Monstera, the new pot should never be more than 2” (or 5cm) larger than the old pot in diameter.   Underwatering may not directly cause root rot. However, when you forgot to water your Monstera for an extended period of time, the root system will adapt to the dry situation. The roots start to shrink and compact in size.  This reaction causes the pot to have rootless soil sections (similar to placing the Monstera on a large pot).  When you finally water the pot, and the soil takes too long to dry, the root may develop root rot. Fungal Infection Root rot is a fungal infection in the soil. There are a few types of fungi that cause root rot, including but not limited to Pythium, Rhizoctonia, and Fusarium.  If these fungi already live in your Monstera’s soil, it is more likely that overwatering will quickly lead to root rot.  Spores from these fungi travel through the air and are carried by insects. If one of your houseplants has fungal root rot, the fungus has likely travelled to other pots of soil in your home. It will not grow or cause damage unless the soil becomes waterlogged.  Cold Climate If you live in a cold climate, your Monstera will slow down its growth in the winter. During this slow growth period, your Monstera will need less water; thus, you don’t need to water it frequently.  Check the soil before you water to ensure that the top two inches (5 cm) are dry to prevent root rot.  Rocks at the bottom of the pot Contrary to popular belief, rocks at the bottom of a plant pot do absolutely nothing to drainage. The only thing they do is keep soil from falling out the drainage hole, which doesn’t usually end up being an issue anyway.  Soil acts like a sponge, soaking up water until it is fully saturated. Only when the lowest soil points are fully saturated does the water start to leak out the bottom. Putting rocks or gravel at the bottom of the pot raises the lowest level of soil, raising the height of that soggy soil.   With rocks at the bottom of the pot, your Monstera is basically in a shallower pot with soil that doesn’t drain any better.  Signs of Root Rot in Monstera Stages of Root RotSigns to observe on your Monstera Early detection• Stagnant growth •Wilted leaves • Yellow leaves Advanced root rot• Root discoloration • Foul odor roots • Black spots on leaves Beyond saving• Roots are all dead • Dark and mushy stems It is crucial to know the signs of root rot in Monstera so that you can prevent an infection from spreading and killing your beautiful swiss cheese plant.  Here are some signs that your Monstera is suffering from root rot, organized by severity. The earlier you notice root rot in your Monstera, the more likely it is to survive!  Early Detection Early signs of root rot in Monstera are stagnant growth, wilted leaves, and yellow leaves. These are all signs that you should check your Monstera’s soil to make sure it is not too wet for too long. With proper care and early detection, your Monstera can easily survive a bout of root rot. Stagnant growth Stagnant growth just means that your Monstera isn’t growing as fast as it usually does. This is a normal thing that happens in the winter when the temperature drops and the days get shorter. Monsteras also slow their growth when they don’t get enough light, fertilizer, or water.  Just because your Monstera is slowing down its growth doesn’t mean its roots are rotting, but it does mean that it could happen if you don’t also slow down its watering schedule.  A plant like Monstera growing slowly will not need as much water, and too much water during stagnant growth can cause root rot.  Wilted leaves Wilted leaves indicate that your Monstera isn’t getting enough water from the soil. If your Monstera is wilting, check its soil to see if it is wet or dry.  If the soil is dry, your Monstera is simply thirsty. It would be best if you gave it a good, long drink of water as usual.  If the soil is already wet and your Monstera is droopy, that means its roots are probably damaged from rot and unable to take up the water that is already there. At this point, you should inspect your Monstera’s roots and look for rot.  Yellow Leaves Yellow Monstera leaves can mean a lot of things, including overwatering and root rot. Yellowing leaves is the next step after drooping leaves if your Monstera continues to sit in soggy soil.  If your Monstera has yellow leaves, you should definitely check its soil and roots for excess moisture and rot.  Advanced Root Rot If the early warning signs are ignored or go unnoticed, your Monstera’s health will continue to decline.  The signs of advanced root rot are black leaf spots, foul-smelling roots, and root discoloration. These signs mean your Monstera needs immediate and drastic care.  It will need a full inspection, and you will have to cut off the dead or dying parts to keep the plant alive.  Root discoloration  Your Monstera’s roots are supposed to be whitish, like most plants. When root rot sets in, the rotting roots turn brown or black, and they get mushy and slimy. You can’t see this unless you dig into the soil or take your Monstera out of its pot to inspect the roots.  When treating your Monstera’s root rot, all dark and discolored roots will need to be cut off; otherwise, the infection will continue to spread.  Foul odor roots Rotting roots stink!  If you put your face up close to your Monstera’s soil, and it smells like something is rotting, something is definitely rotting, and it’s probably your Monstera’s roots.  The stronger the smell, the larger the infection and the harder it will be to keep your Monstera alive at this point.  Black spots on leaves Black spots will start to show up on your Monstera’s leaves, sometimes accompanied by yellowing around the spot. This is an indication that the infection has started to spread throughout the plant.  Root rot will eventually spread through the entire plant, up the roots to the stem to the leaves. Black or brown spots on the leaves are how you know that process is occurring.  Beyond Saving  Unfortunately, sometimes we miss the signs of root rot until it is too late. Once the rot kills all the roots and moves up into the stem, your Monstera is on its way out.  At this point, the only way to salvage your Monstera would be to try to propagate uninfected sections of the stem, if there are any left.  Roots all dead Your Monstera is unlikely to survive once all its roots are dead. Even though we don’t see them, roots play a vital part in the life of every plant. No roots mean your Monstera can’t take in any water or nutrients, and it will likely die of dehydration.  Dark and Mushy stem Once the rot travels into your Monstera’s main stem, that stem will become dark and mushy like the roots. The rotten stem is unable to transport water and nutrients to the leaves, and the Monstera will soon die.  Inspecting for Root Rot If you notice any of the above signs in your Monstera, you should inspect its soil and roots for rot. If the signs are mild, you can get away with digging away a portion of the soil to look at the roots.  If your Monstera displays signs of more advanced root rot, such as black, slimy roots or black spots on the leaves, you should take it out of its pot for a more detailed inspection.  Taking your Monstera out of its pot for a root inspection is the best way to detect root rot. Unfortunately, it tends to stress your Monstera out, causing it to droop its leaves for a little while. It is worth it to inspect because a little bit of stress on your plant is not as bad as letting its roots rot.  When you inspect your Monsteras roots for rot, look out for the following: • Brown or black slimy roots that break easily • Waterlogged soil • Foul-smelling roots and soil • Mushy roots and stem If you see these signs of root rot in your Monstera, proceed to the next section to learn how to treat it!  Treating Root Rot in Monstera Now you know if your Monstera has root rot, it is time to learn how to treat it so your Monstera can keep living a healthy life. Step 1: Rinse the roots. Since root rot is a disease that lives in the roots and soil, we need to get rid of all the infected soil. Take your Monstera out of its pot and brush away as much soil as you can with your hands. If the rot is nasty, some roots will come off easily with the soil.  Once you get rid of the soil, you will get a better idea of how far the rot has eaten into your Monstera.  You need to get rid of all the infected soil. Once you are done brushing it away manually, rinse off the rest with water. A small Monstera can probably fit in the kitchen sink, but a large specimen will need to be sprayed down with a hose outside. Step 2: Cut the damaged roots Once all the soil is gone, you need to cut off all the dead and rotting roots.  You do not want to spread the fungal infection to healthy roots! Use rubbing alcohol or diluted bleach to disinfect your pruners in between every single cut.  Cutaway as much infected material as you can from your Monstera. You don’t want to leave any behind that can grow into another significant infection.  If, by the time you’re done cutting away the dead parts, there are no more roots or the rot extended into the stem, skip to Step 7.  Step 3: Let the roots air-dry, disinfect roots and pot. Just in case there is still rot lingering on the roots, you need to let them dry out before repotting. For extra disinfection, you can also dip the remaining roots into diluted hydrogen peroxide (1 tablespoon peroxide for each cup of water) before setting them out to dry.  While the roots sit out to dry, prepare its pot and new soil. Clean the pot, and any supports completely and disinfect with diluted bleach or peroxide. Dispose of the old, contaminated soil. Step 4: Repot in fresh soil. Once your Monstera’s healthy roots have dried out enough, it is time to repot! You must use fresh, new soil for your Monstera. Old, infected soil will just cause your plant to develop root rot again.  Now is the time to adjust your potting soil to have better drainage by adding more perlite or sand to the mix. Better draining soil will help prevent your Monstera from developing root rot in the future. Step 5: Prune affected leaves It is a good practice to cut back a third to a half of your Monstera’s leaves when you cut back its roots. With fewer roots, it won’t be able to support as many leaves, and they will droop. Cutting back leaves affected by root rot will allow your Monstera to bounce back faster. Step 6: Place in bright sunlight Monsteras are happiest in bright, indirect light. When they don’t get enough light, they don’t take up as much water and are susceptible to developing root rot.  Not only will bright light keep your Monstera from developing rot, but it will also help it grow bigger and taller. Also read: How Much Light Does a Monstera Need? Step 7: (In severe cases) Propagate from healthy cutting. If, when cutting back the roots, you find that your Monstera has no healthy roots left, you won’t be able to repot it as normal. At this point, you will have to help it grow new roots by propagating stem pieces.  First, cut off all rotten parts of your Monstera—roots, stems, leaves, anything that has been affected by root rot.  Do you have any pieces of stem left with leaf nodes? If so, great! Your Monstera is not dead yet!  There are two main ways to propagate Monstera stem cuttings: you can plant them as if you were repotting them, or you can propagate them in water.  To propagate them in soil, just follow steps 4 and 5 above. Avoid placing the baby Monstera in bright light until it has started to develop roots. You can check to see if it has roots yet by giving it a light tug.  For water propagation, place the Monstera stem in a glass or vase of clean water. Change the water every few days to keep it clean. Once it starts growing roots, you can repot it as in Step 4. How to Prevent Rot Root Ultimately, the best way to keep your Monstera from dying from root rot is to prevent it from happening in the first place. Thankfully this is easy to do with a little bit of extra attention to your amazing plant! Proper Drainage Your Monstera’s pot absolutely must have drainage holes. There is no way to get around this.  Unfortunately, Most of the prettiest pots do not have drainage holes. You can remedy this in one of two ways: drill a drainage hole in the pretty pot or use a simple plastic pot with drainage inside the pretty one. Also Read: How to Choose the Perfect Pot for your Monstera Aerate the soil  In addition to proper drainage, your Monstera’s soil needs aeration.  Adding perlite or sand to your potting mix during repotting will allow the water to flow more freely through the soil and allow more oxygen to reach your Monstera’s roots Also read: Choosing the Best Soil Mix for Monstera to Thrive. Water Schedule Even if your Monstera has adequate drainage and soil, it may still be waterlogged if you are watering it too frequently.  Your Monstera’s watering schedule may need to be longer if it does not get as much sunlight or if its growth slows down like during the winter.  This leads to my next point… Check Soil before Watering The only way to know for sure if your Monstera is ready to be watered or not is to check the soil before watering. Your watering schedule should only be a guideline, and you should always adjust it based on your Monstera’s actual needs.  To check the soil, stick your finger straight down into the top layer. If the soil is still damp, you don’t need to water yet. Your Monstera is ready for water once the top 2-3 inches (5-6 cm) are dry. Inspect for Pest & Fungal Infection Regular inspections of your Monstera will help you catch infections before they become deadly. Especially during the winter, you should make a point of regularly inspecting your Monstera’s leaves, stems, and soil for signs of pest and fungal infections.  Signs of infections include drooping leaves, yellow leaves, and soil discoloration (in the case of a fungus).  Natural Fungicide for fungal Black Root Rot In the case of fungal root rot, there are natural fungicides you can use to help treat your Monstera.  These include  Hydrogen Peroxide Hydrogen peroxide is a bleach and disinfectant which, when diluted, can be used to treat root rot.  To use hydrogen peroxide to treat root rot in Monstera, mix one tablespoon of 3% strength peroxide and one cup of water. This can be used directly on the roots after pruning or as a supplemental oxidization treatment during watering.  The extra oxygen molecule in hydrogen peroxide fights directly against the anaerobic environment in which root rot thrives! Cinnamon is a natural fungicide and rooting agent.  Use cinnamon when treating your Monstera for root rot by dusting it on its freshly-cut roots (or stem). The cinnamon will help prevent rot from affecting the healthy tissue, and it will stimulate new root growth. Baking Soda Baking soda, or sodium bicarbonate, is another natural and organic fungicide used to treat root rot.  Be careful not to use too much too often, as baking soda is a salt, and too much salt of any kind is not good for your Monstera.  To use baking soda as a fungicide, mix three tablespoons into a gallon of warm water. Spray it into your Monstera’s soil and stems in the evening to help prevent a fungal root rot infection. Apple Cider Vinegar You can also use apple cider vinegar to combat fungal black root rot in your Monstera. Like the baking soda, mix ACV with a gallon of water and spray onto your Monstera’s leaves, stem, and soil. Neem Oil Neem oil is excellent for getting rid of insect and fungal infections in your Monstera. Neem oil can be used to combat fungal root rot when used as a soil drench.  Mix a teaspoon of neem oil with a quart (1 liter) of water. Add a couple of drops of dish soap to emulsify the mixture, and mix well. Apply this drench to your Monstera’s soil after watering. Related Questions: Can you reuse soil that has root rot? Absolutely not!  Root rot causing fungi can live in the soil for years, and all it takes is a few days of too much water for them to start attacking your Monstera’s roots again. Infected soil should be disposed of.  Do not compost diseased plants or soil, as this will only lead to the infection possibly spreading.  Gabriella Anastasia I have been growing houseplants for most of my life and have been collecting them ever since I was 11 years old.  Now 33, I've always had a green thumb (even though my parents thought otherwise) and love to share my knowledge with others. Recent Posts
What is the Catholic Conference? What is the Pennsylvania Catholic Conference? Congress is about to vote on the most radical abortion bill of all time. 1975) would impose abortion on demand, at any stage of pregnancy. … It would apply nationwide—even in states that don’t want it. What are the six divisions in the Catholic Church? What are the six divisions in the Catholic Church? In these two divisions there are six branches: Catholicism, Protestantism, Eastern Orthodoxy, Anglicanism, Oriental Orthodoxy, and Assyrians. Restorationism is sometimes considered the seventh branch. Who is above the Pope? Cardinal. Cardinals are leading bishops and members of the College of Cardinals. Their biggest duty is participating in the Papal Conclave, that is, voting for the new Pope. Most have additional duties including missions within the Roman Curia, governmental body of the Holy See. What is the difference between an archbishop and a cardinal? What was the first religious order? In particular the earliest orders include the English Benedictine Confederation (1216) and Benedictine communities connected to Cluny Abbey, the Benedictine reform movement of Cistercians, and the Norbertine Order of Premonstratensians (1221). THIS IS IMPORTANT:  How old is pastor John Wesley? Who can celebrate all 7 Sacraments? Bishops are said to have the “fullness of the priesthood,” because they alone have the authority to offer all seven sacraments — Baptism, Penance, Holy Eucharist, Confirmation, Matrimony, Anointing of the Sick, and Holy Orders.
Decisions, visions, brain functions and storytelling Recently, I’ve become interested in decision making. My job is frequently about helping people make decisions which impact their businesses and their lives. It’s also about selling services, which requires some clue about how and why people make purchasing decisions. And these professional interests in decision making, and underscored by the constant stream of decisions contemplated every day, both large and small. What shirt will I wear? What car should I buy? Where should my kids go to school? With all this bouncing around my prefrontal cortex, I’ve found much insight in Jonah Lehrer’s book How We Decide. It’s about what happens in our brains when we’re making decisions, the roles played by rationality and instinct. It’s also about which parts of our brains are used when making these decisions. As you’d expect for a popular science book, narrative accounts play an important role in bringing the various examples of decision making to life. When you’re kicking around terms like ‘stochasticity’, ‘posterior cingulate’ and ‘dopamine receptors’, it helps if you can relate it to stories about football matches and Deal or no Deal. There’s a couple of stories about mid-flight incidents on board passenger jets, a topic which can always be relied upon to raise the heart rate. One is about a United Airlines flight from Denver to Chicago, which was interrupted by an internal explosion which took out all the hydraulics, leaving the pilots without the ability to steer the plan. The story is about decision making under pressure, and the ability to invent a tactic for landing the plane on the run. “…[the pilot] needed to solve his problem, to invent a completely new method of flight control. This is where the prefrontal cortex really demonstrates its unique strengths. It’s the only brain region able to take an abstract principle – in this case, the physics of engine thrust – and apply it in an unfamiliar context to come up with an entirely original solution. It’s what allowed [the Pilot] to logically analyse the situation, to imagine his engines straightening [the plane’s] steep bank.” This last phrase, about imagining an end result and creating a process to achieve that result, caught my eye. This is, I think, what many creatives do. They imagine the end result and corral the resources (time, capital, labour etc) to bring about that vision. Some will be able to design prototypes to communicate that vision to others in the production process; a fashion designer will do so through sketches and patterns. Others will have to do so without visual tools; a screenwriter has to imagine what a film will look like and sell that vision to others, long before a frame is shot. Entrepreneurs do this too. They have to imagine a version of their business which fulfils what they want from it: money, lifestyle or whatever it is that sparks their motivation for being in business. They have to imagine the end result and ‘see’ it long before others can. Then they invent a way to achieve it. And ‘invent’ is really the right word because although they can follow the steps others have taken in the past, each business’s journey is unique, with its own ups and downs. The Pilot’s story is about the suppression of emotion (in this case, panic) to focus on rationality.( It’s not always like this though; the book also highlights decision making which is enhanced by instinct and emotion). But it’s also about the ability to screen out all unnecessary information in order to concentrate on the crucial data. For example, the Pilot had no time to focus on the hydraulics – they were gone and never coming back. He had to focus on the elements he could control – in this case engine speed –and block out the rest. Again, there’s something in this for entrepreneurs. I met this week with a fellow who runs a creative industries business and he told me that his company now focuses less on small, labour intensive jobs and more on larger scale jobs for corporate clients, on which he can spread his resources more evenly. Such is the dream of many a small business owner, so I asked him how he achieved it. He didn’t really know how; there had been no deliberate strategy, other than to adopt a vision for his company which involved work with large corporate clients. He was inspired to do so by a presentation by a Hindu priest he met at a bank’s innovation conference (I know, right?). The priest talked about balancing a peacock feather on your finger. If you look at your finger, apparently it’s really hard to balance the feather. But look at the top of the feather, and it’s much easier. (Peacocks are hard to come by at my place, so I have yet to try this for myself.) Choose the metaphor you like – looking at the top of the feather, forgetting about the plane’s hydraulics – the point is that focus on an end goal and screening out distractions count for something. And that there’s nothing like telling a good story to illustrate abstract concepts. PS. Talking of good stories..While adding some links to this article, I discovered that How We Decide has been withdrawn from sale by its publisher. The story’s here. Ref: Lehreh, J. 2009, How We Decide. Houghton Mifflin Harcourt, New York. Decision making and electing entrepreneurs
On the Origins of Cancer in the Paradox of Good Health ImageA recent New York Times article, In Sickness and in Health: a Wedding in the Shadows of Cancer, struck a particularly sensitive chord. It profiled the courageous journey of two of my patients, a husband with advanced pancreatic cancer and his wife with metastatic breast cancer and their preparation for the wedding of their daughter in the face of their battle with cancer. Among the responses of readers, an interesting and common concern stood out, why we continue to have an epidemic of cancer and why have we not dealt with the toxic mix of chemicals in our environment that they assume are a major cause of these cancers. This perception is widespread. While there is little doubt that there are adverse health effects from these exposures, their role in the origin of the typical “western cancers” (breast, colorectal, prostate, etc.) remains uncertain. Certainly the central role of tobacco in lung cancer and several other cancers is undisputed. What is most striking is the lack of awareness among the public and health care providers alike about the growing evidence that many non-smoking related cancers may actually be, in part, a direct effect of modern good health.  The marked improvement in maternal and early childhood nutrition has played an important role in the health and well being of modern populations. The seminal work “The Changing Body”(1) by the late Nobel laureate in Economics, Robert Fogel and colleagues, detailed the dramatic improvement in health, nutrition and human development in the Western World since the 1700’s. The single best indicator of improved population health is adult height.  Indeed, increasing height is linked to improved longevity, decreased mortality from heart disease and improved socioeconomic status (2). Perhaps the single most important benefit has been the significant decrease in early childhood mortality, declining from 30 to 50 % before age 10 prior to the 20th Century to under two percent now. While maximal height is in part an inherited trait, our attained height is largely determined by both pre- and postnatal influences, including maternal weight and diet as well as early childhood calorie and protein intake. Why do I suggest a direct effect of good health on the origins of cancer?  One of the most consistent associations with cancer risk, recognized over the last 25 years (3-6), is the increased risk of cancer with increasing adult height.  A recent study, published in the Journal of the National Cancer Institute (7) suggested that height may in large part explain the higher risk of cancer in men compared with women.  It has been speculated that increased height associated with earlier puberty may be a potential explanation for some hormone related cancers such as breast (8). The height-cancer association, in fact, extends well beyond these cancers and appears to be a marker of cancer risk in general. Many non-hormonal cancers such as melanoma, lymphoma and leukemia share this association (9). That something as simple as one’s height could contribute to a disease that occurs 50 or 60 years later seems difficult to understand at first. But, we now know that late adult cancers have their origins decades prior to diagnosis. The earliest event in cancer is a random mutation in a critical gene of a specific cell, the tissue stem cell (10). Within each organ, the tissue stem cells represent a tiny fraction of all the cells present. When these stem cells divide, one cell may grow to provide new tissue cells and one remains behind as a stem cell, allowing for continued “self renewal”. These stem cells provide a source of new cells that maintain the viability of the tissues over the life of the individual. This critical “initiating” mutation is passed to that cell’s progeny.  Over time, additional mutations may occur in additional critical genes that are necessary for the emergence of a fully malignant cancer. There are hundreds, if not thousands of mutations within the cells in a newly diagnosed cancer, only some of which may play a crucial role in the malignant process (11). While exogenous carcinogens (tobacco, organic chemicals, radiation) are in part responsible for these mutations, many result from endogenously generated radicals as an inherent part of cellular metabolism (12). How does height influence cancer risk? Height does not cause cancer. But what height does reflect is an increased numbers of cells in every organ. Robust prenatal and early postnatal growth are accompanied by an increase in the number of tissue stem cells in each organ, each vulnerable to that initiating mutation, in what I have called the “lottery ticket effect” or pre-initiation. This has been referred to as “the stem cell burden hypothesis” (13), though it may be less a burden than a natural consequence of robust early growth. The central hormone involved in fetal growth and a major determinant of childhood and later adult height is Insulin-like growth factor -1(IGF-1).  Dimitri Trichopoulos and colleagues, leaders in the research linking height to cancer, showed that IGF-1 levels are closely correlated with specific stem cells (blood-related (hematopoietic) and breast tissue stem cells) when measured at birth in umbilical cord blood (14) and those stem cells are closely linked to birth length, an important predictor of adult height. That IGF-1 would play such a critical role is reflected in the abundant research in animal models of cancer, indicating that calorie restriction markedly reduces cancer development, as a result of lower  IGF-1 levels (15).  In the large body of research in humans, increased IGF-1 levels are associated with the risk of many cancers (16). Thus, the beginnings of the modern cancer epidemic may well start in these early events that ironically reflect an early environment promoting vigorous childhood growth (17). Later events including increased exposure to exogenous carcinogens, the continued exposure to this same rich caloric environment of modern life and our sedentary lifestyle, provide the fuel necessary to accelerate the evolution of the cancer process, resulting in the high cancer burden we now experience. We cannot and should not want to limit the early and healthy growth of children. But we can target these later factors that ultimately are at the heart of the cancer epidemic. I invite you to ‘like’ my Facebook page, follow me on Twitter, join my LinkedIn network, and please visit my website, www.drbarryboyd.com. 1. The Changing Body. Health, Nutrition and Human Development in the Western World since 1700. Roderick FLoud,  Robert  Fogel, Bernard Harris and Sok Chul Hong. Cambridge University Press, Cambridge, UK, 2011. 2. Height, wealth, and health: an overview with new data from three longitudinal studies. Batty GDShipley MJGunnell DHuxley RKivimaki MWoodward MLee CMSmith GDEcon Hum Biol. 2009 Jul;7 (2):137-52. 3. Adult stature and risk of cancer at different anatomic sites in a cohort of postmenopausal women. Kabat GCAnderson MLHeo MHosgood HD 3rd,et al. Cancer Epidemiol Biomarkers Prev. 2013 Aug;22(8):1353-63. 4. Height and cancer incidence in the Million Women Study: prospective cohort, and meta-analysis of prospective studies of height and total cancer risk.Green J, Cairns BJ, Casabonne D, Wright FL, Reeves G, Beral V; Million Women Study collaborators.Lancet Oncol. 2011 Aug;12(8):785-94. 5. Adult height in relation to mortality from 14 cancer sites in men in London (UK): evidence from the original Whitehall study. Batty GD, Shipley MJ, Langenberg C, Marmot MG, Davey Smith G. Ann Oncol. 2006 Jan;17(1):157-66. 6. Height and site-specific cancer risk: A cohort study of a korean adult population.Sung JSong YMLawlor DASmith GDEbrahim SAm J Epidemiol. 2009 Jul 1;170(1):53-64. 7. Height as an explanatory factor for sex differences in human cancer.Walter RB, Brasky TM, Buckley SA, Potter JD, White E. J Natl Cancer Inst. 2013 Jun 19;105(12):860-8. 8. Body measurements, estrogen availability and the risk of human breast cancer: a case-control study. Bruning PF,  Bonfrèr JMHart AAvan Noord PA, et.al. Int J Cancer. 1992 Apr 22; 51(1):14-9. 9. Adult height and the risk of cause-specific death and vascular morbidity in 1 million people: individual participant meta-analysis. Emerging Risk Factors Collaboration Collaborators. Int J Epidemiol. 2012 Oct;41 (5):1419-33. 10. A mathematical model of cancer stem cell driven tumor initiation: implications of niche size and loss of homeostatic regulatory mechanisms. Gentry SNJackson TLPLoS  One. 2013 Aug 19; 8(8):e71128. 11. Human cancers express mutator phenotypes: origin, consequences and targeting. Loeb L ANat Rev Cancer. 2011 Jun;11(6):450-7. 12. Links between metabolism and cancer. Dang CVGenes Dev. 2012 May 1; 26(9):877-90. 13. Association of fetal hormone levels with stem cell potential: evidence for early life roots of human cancer. Baik I, Devito WJ, Ballen K, Becker PS, et. al.Cancer Res. 2005 Jan 1;65(1):358-63. 14. Correlation of umbilical cord blood haematopoietic stem and progenitor cell levels with birth weight: implications for a prenatal influence on cancer risk. Strohsnitter WCSavarese TMLow HPChelmow DP, et. al. Br J Cancer. 2008 Feb 12;98(3):660-3. 15. Roles for insulin-like growth factor-1 in mediating the anti-carcinogenic effects of caloric restriction. Kari FWDunn SEFrench JEBarrett JCJ Nutr Health Aging. 1999;3(2):92-101. 16. The insulin-like growth factor system in cancer. Weroha SJ, Haluska P. Endocrinol Metab Clin North Am. 2012 Jun; 41(2):335-50. 17. Are cell number and cell proliferation risk factors for cancer? Albanes D, Winick MJ Natl Cancer Inst 188; 80:772-74. Leave a Reply WordPress.com Logo Google photo Twitter picture Facebook photo Connecting to %s
Recycling And Reuse Of Cardboard Many people don’t know that cardboard can actually be recycled and reused to make new products. Recycling old cardboard boxes is great for the environment because it keeps your trash out of landfills and reduces the need to cut down trees for paper products. Cardboard recycling can also create jobs which helps boost the economy. For more information please visit on website or call on : +91 9820020714 Recycling and reuse of Cardboard A common misconception about recycling is that it takes too much time and energy to separate an object into its component parts. The fact is, recycling can save you time, saved fuel, and money! Cardboard is an extremely versatile material that can be recycled for various purposes. Cardboard boxes, for example, can be reused as a container for storing items such as food and mail. Furthermore, cardboard boxes are a great way to store clothes. If you need a sturdy yet inexpensive box to ship or store something, then a previously used cardboard box should suit your needs. Simply remove tape and labels from the boxes before placing them upside down on flat surfaces. Cardboard Shredder For Packaging Material That Saves A Lot Of Money Cardboard shredder for packaging material that saves a lot of money. This article will explore how a cardboard shredder can help save a lot of money through recycling and reducing the amount of waste sent to landfills. For starters, the company’s shredding plant handles about 10 tons an hour. Their process can handle up to 20 inches of cardboard at one time, each batch lasting only 5 minutes. Cardboard Shredder for packaging material Cardboard shredder for packaging material is a great way to save money because it recycles cardboard. Cardboard is a very common and cheap material, which means that this machine can create a significant amount of savings. It also prevents the spread of bacteria since cardboard debris often carry germs from food products. This machine is also able to cut down on landfill waste as it reduces cardboard by as much as 80%. To save money for companies, some are opting to use cardboard shredder machines that will cut down on the amount of cardboard needed. The shredders can convert up to 3 inches of cardboard into just under 1 inch. This means that the cardboard is used more than twice as much by cutting it in half, which saves money on storage and also on shipping cost. If you would like more information please contact us.
Avoid These 10 Things That Can Trigger Insomnia Insomnia is a type of sleep disorder that does not let the patient fall asleep even if he wants to. Around 30% of adults are going through temporary or acute insomnia and 10% are suffering from a long-term sleep disorder. People who believe in myths for curing insomnia usually trigger it even more. This is why everyone should be well-informed about such issues, and that’s exactly what this article is about. Let’s dive into details: Avoid These 10 Things That Can Trigger Insomnia Staying Out Late When people stay out late even after completing 8 to 10 hours of work, they are more likely to suffer insomnia. Even if you sleep, you won’t be able to take a long nap. When you reach home late at night, your brain may not produce melatonin during those hours. So, it would be better if you quit this habit and reach home after office. A comfortable routine is important to sleep properly. According to doctors, people should try to sleep at 10:00 pm in the night because this is the best time when the brain starts producing melatonin. Taking Too Much Stress If you are not good at stress management, you may suffer from insomnia. Loss in business, poor academic performance or relationship problems lead to stress. People usually go through acute insomnia due to stress and if they fail to get back to a normal routine, long-term insomnia is possible. Panic attacks and PTSD are also reasons for chronic insomnia. It is important to learn basic tips to avoid stress in your life. Irregular Sleep Time Irregular sleep time is another major reason for insomnia because people do not build a certain routine to fall asleep. If you sleep at noon for about 4 hours, you may not feel sleepy at night. Those who wake up late in the morning also face trouble while sleeping at night. It is important to improve your routine instead of irregular patterns. Time management is vital. You should prefer to do office work in the daytime whereas a short nap of thirty minutes in the noon is good enough. Mental Illness If you suffer different mental illness problems, the chances of acute or chronic insomnia are high. People with depression issues cannot sleep peacefully. Bipolar disorder, obsessive-compulsive disorder, and anxiety also trigger insomnia in a person. If you are going through such problems, we suggest visiting a psychiatrist who can suggest you medication to get over it. Mental illness is curable and once you successfully get over it, you will not suffer insomnia. Breathing Problems Asthma may cause insomnia to some people who have severe breathing problems. Heavy snoring can be sleep apnea and you may not remember it at the moment but after waking up the next morning, you may feel groggy. Nasal allergy can also cause this problem. Apart from that, people who gain excessive weight should also stay careful because it is also a reason for insomnia. You must visit a specialist to discuss breathing problems. Consider buying an air purifier for your house. There are also many air purifying indoor plants that can do the same for you. This will ensure that you keep getting clean and fresh oxygen while you sleep. For example, Peace Lily and Snake Plant are some asthma-friendly houseplants. Uncomfortable Bedroom People who sleep on an uncomfortable bed also go through trouble in the night because neck and back pain do not let them sleep comfortably. The first important thing is to bring a new bed in the house with a comfy mattress. The bed frame should be adjustable to different positions whereas mattress sizes are also vital to consider. We suggest choosing a mattress that is medically proven for giving a smooth touch to the back.  You should prefer to improve the overall environment of the bedroom by reducing the brightness of lights at night. Once you design a cozy bedroom, the environment will prove helpful to sleep peacefully. Memory loss further leads to multiple problems. Dementia and Alzheimer’s are also reasons for dementia. Most people with dementia get restless after sunset which is often called sundown syndrome. They get confused and panic around bedtime and start roaming in the house instead of going to bed. The majority of people go through this situation till late at night and then they fall asleep whereas a few stay awake the whole night. It depends on the type of dementia and a person’s immune system how he copes with it. Some medicines also leave adverse effects in the form of acute insomnia however when you leave that medication, insomnia can also be cured. So, if you are taking some medication right now, we suggest consulting with a doctor regarding the issue so he may make some changes to that. Acute insomnia is easily curable so you should not be worried about it. Remember that each medicine comes with side effects. Try to look for alternative natural methods to improve your condition if possible. Only take medicines as a last resort and don’t consider them your friend. What May Help You Sleep? Sometimes, people think that certain types of drinks or drugs can help them sleep. However, they are not appropriate choices because drugs can even trigger the problem after providing temporary relief. You should avoid overthinking and managing stress can help a lot. Those who have gained a lot of weight should go for exercise regularly. It is vital to avoid sleep pills because they are not effective for curing insomnia. A peaceful environment is also vital to fall asleep comfortably so you must design the bedroom carefully. The above-mentioned points clearly state the reasons for insomnia. It is important to consult a doctor without an unnecessary delay. When people try to find a cure for insomnia on their own, they make several mistakes. You should go for therapy sessions if something is causing you depression because consultation is important. In short, people should stay careful about their daily routine because small things affect their health and sleeping patterns. It would be great if you come home early and try to sleep on time. Leave a Comment
Threaded Fasteners: A Quick Guide hexagon bolts Threaded Fasteners: A Quick Guide As we move into the digital age, information technology is changing all industries. From robotics and communications to manufacturing, every sector is undergoing rapid transformation as we make our way through what is known as the 4th industrial revolution. However, this seismic change is being underpinned by the smallest of components, such as threaded fasteners. Machines simply cannot function without them. Industrial machines, electrical appliances, and electric vehicles are just a few of the devices that rely upon threaded fasteners to join and secure their disparate elements. In fact, without these threaded mechanical fasteners, none of this current digital revolution can take place. So with that in mind, let’s walk you through threaded fasteners and bring you up to speed with their role in today’s economy. What Are Threaded Fasteners? In short, a threaded fastener uses a screw head thread to advance within a nut or pre-tapped hole to enable the connection of two (or more) parts. The fact that these fasteners are threaded means that they can be disassembled and used again if necessary. For this reason, they are considered non-permanent fasteners. Another important attribute that distinguishes threaded fasteners from non-threaded fasteners is their strength. The grooves cut into the material actually enhance the mechanical strength of the material (after the hardening process), making it stronger than non-threaded counterparts. Threaded Fasteners Applications: Who Uses Them? Just about everyone! Whether you realise it or not, threaded fasteners are present in everyone’s daily lives. From the screws holding your sofa together from the industrial machines that manufactured the smartphone in your hand, they are all held together by threaded fasteners. That’s why, as mentioned, they are playing a vital role in our shift towards a digital economy. AI-driven machinery, robotics, IoT-enabled devices, 5G networks, they all rely upon threaded fasteners in one way or another. They also form the bedrock to more traditional industries such as the maritime or furniture sector. In essence, they touch our lives every day without us ever taking the time to realise it. What Are Some of the Most Common Types of Threaded Fasteners? There are many types of threaded fasteners. However, they tend to fall under two broad categories: bolts and screws. These fasteners are used in conjunction with a nut. They have a thread on one end and a head on another. There are several bolt head types, with the most common being round, countersunk, hex, and flat. Nuts are different shapes, too, with the most common being square or hex. Screws are very similar to bolts in that they also have a head on one end with the thread on the opposite end. The only difference arises from the length of the thread, with screws usually having longer threads than bolts. The head is usually manufactured to be compatible with a screwdriver or spanner, facilitating a firm grip when driving the screw in. Where Can You Source Threaded Fasteners? While generic threaded fasteners are available at national and local hardware retailers, you’ll need to approach a specialist such as Leyton Fasteners when looking for specific fasteners for a specific function. We carry every type of threaded fastener imaginable, including obsolete fasteners that are no longer commercially produced. From maritime threaded fasteners that hold some of the world’s biggest boats together to tiny intricate screws for classic cars, we stock anything you might require for your business or personal endeavours. We offer next-day delivery nationwide, and our team is always on hand to answer any queries you might have regarding our threaded fasteners on 0151 355 8045. So if you think you are in need of a batch of threaded fasteners, make sure to browse our online shop or visit us in person. Our opening hours are Mon-Thur: 8:30 – 17:00 and Fri: 8:30-16:00, and we look forward to welcoming you to our premises soon.
The reasons why the Cold War between the USA, the Soviet Union and other participants started were complex and hard to identify independently. The main reason was the ideological difference between the two models of society. The first one was socialistic and the second one was capitalistic. The causes of the Cold War included the struggle for the spheres of influence of the USA and the Soviet Union. Both sides wanted to expand their powers on other territories. The USA and the USSR supported different regimes and in some places, it turned into the local wars. Among the other reasons, scientists mention the refuse of the Soviet Union to become a part of the UN, the support of the allies of western countries of the Atlantic Charter in the year 1941, the mutual fear of the USSR and the USA of their nuclear weapons. Also one of the reasons for the Cold War was the Soviet wish to promote and spread communism all over the world. The American reaction on the threat of communism was the creation of NATO in 1949. Even before that event, the USA decided to restrain the Soviet Union after their shown motives in Greece and Turkey where the USSR wanted to have their military bases. The Truman Doctrine finally ended the previous cooperation between the countries in 1947. The USA was involved in the Korean and Vietnam wars because of their fear that the communistic regime will get strategically useful places there after its spread all of the countries’ territories. The Korean Peninsula happened to be separated into two parts and two countries. The Soviet Union together with China supported North Korea and the USA supported South Korea. The participation of the USA in the Korean war was another stage of the confrontation with the communistic ideology of the USSR. The same was during the Vietnam war where the USA fought against the communist principles which were being planned to grow there after the achievement of independence. Americans supported South Vietnam in that war. By the year 1966 more than two hundred thousand troops were sent to Vietnam from the side of the USA. You're lucky! Use promo "samples20" and get a custom paper on "American Involvement in the Cold War" with 20% discount! Order Now
Are receptors integral membrane proteins or peripheral membrane proteins? Integral membranes proteins serve as transporters. Peripheral proteins serve as cell adhesion molecules, antigens and enzymes. So what about receptors? Which proteins carry out the duty of receptors? All the receptors I know about are integral, transmembrane proteins. It would certainly be possible for a receptor domain to exist on a peripheral protein that interacted with a transmembrane protein, but I don't know of any examples of this: an extracellular peripheral protein would tend to float away, which is a bit of a problem if they aren't intended for secretion. There are certainly peripheral proteins that associate with the internal parts of transmembrane receptor proteins and may influence their function, but they are not themselves receptors. I'm not sure about your question about "which proteins" - those proteins are most often named after their ligand followed by the word "receptor", or based on their function that is triggered by binding their ligand, or both (i.e., the AMPA receptor is a glutamate receptor that is also activated by AMPA; receptor tyrosine kinases are a family of receptors that are tyrosine kinases; the individual members of the family are typically known by their ligand). Lipid-anchored protein Lipid-anchored proteins (also known as lipid-linked proteins) are proteins located on the surface of the cell membrane that are covalently attached to lipids embedded within the cell membrane. These proteins insert and assume a place in the bilayer structure of the membrane alongside the similar fatty acid tails. The lipid-anchored protein can be located on either side of the cell membrane. Thus, the lipid serves to anchor the protein to the cell membrane. [1] [2] They are a type of proteolipids. The lipid groups play a role in protein interaction and can contribute to the function of the protein to which it is attached. [2] Furthermore, the lipid serves as a mediator of membrane associations or as a determinant for specific protein-protein interactions. [3] For example, lipid groups can play an important role in increasing molecular hydrophobicity. This allows for the interaction of proteins with cellular membranes and protein domains. [4] In a dynamic role, lipidation can sequester a protein away from its substrate to inactivate the protein and then activate it by substrate presentation. Overall, there are three main types of lipid-anchored proteins which include prenylated proteins, fatty acylated proteins and glycosylphosphatidylinositol-linked proteins (GPI). [2] [5] A protein can have multiple lipid groups covalently attached to it, but the site where the lipid binds to the protein depends both on the lipid group and protein. [2] Structure of Peripheral Proteins In the image below, several peripheral proteins are labeled. A peripheral protein does not have a definite structure, but it has several key aspects which make it a peripheral protein. First, all peripheral proteins are associated with the cell membrane. The amino acid sequences of these proteins are unique in that they draw the proteins to the membrane, and they tend to congregate on the surface of the membrane. This allows them to be in the right place to carry out their designated action. In the image, the orange peripheral proteins are seen attached to either the phosphoglyceride lipid molecules which make up the lipid bilayer, or to integral proteins. A protein without these areas of amino acids would not be attracted to the membrane. It would be distributed evenly throughout the cytoplasm, and would not be a peripheral protein. Second, peripheral proteins do not have a hydrophobic region of amino acids. This, and the polarity of other amino acid groups, keeps the peripheral proteins on the surface of the cell membrane. This is due to the amphipathic nature of phosphoglycerides. This means that the blue “head” region is polar and hydrophilic. The yellow “tails”, which constitute the middle of the membrane, are hydrophobic. To avoid being sucked into the membrane, peripheral proteins often have lots of hydrophilic amino acids exposed on their surface. Integral proteins expose hydrophobic amino acids in the middle, and hydrophilic amino acids on the parts exposed to water. This effectively locks them within the membrane. The plasma membrane contains molecules other than phospholipids, primarily other lipids and proteins. The green molecules in the figure below, for example, are the lipid cholesterol. Molecules of cholesterol help the plasma membrane keep its shape. Many of the proteins in the plasma membrane assist other substances in crossing the membrane. The plasma membranes also contain certain types of proteins. A membrane protein is a protein molecule that is attached to, or associated with, the membrane of a cell or an organelle. Membrane proteins can be put into two groups based on how the protein is associated with the membrane. Integral membrane proteins are permanently embedded within the plasma membrane. They have a range of important functions. Such functions include channeling or transporting molecules across the membrane. Other integral proteins act as cell receptors. Integral membrane proteins can be classified according to their relationship with the bilayer: • Transmembrane proteins span the entire plasma membrane. Transmembrane proteins are found in all types of biological membranes. Some integral membrane proteins are responsible for cell adhesion (sticking of a cell to another cell or surface). On the outside of cell membranes and attached to some of the proteins are carbohydrate chains that act as labels that identify the cell type. Shown in the figure below are two different types of membrane proteins and associated molecules. Peripheral membrane proteins are proteins that are only temporarily associated with the membrane. They can be easily removed, which allows them to be involved in cell signaling. Peripheral proteins can also be attached to integral membrane proteins, or they can stick into a small portion of the lipid bilayer by themselves. Peripheral membrane proteins are often associated with ion channels and transmembrane receptors. Most peripheral membrane proteins are hydrophilic. Some of the membrane proteins make up a major transport system that moves molecules and ions through the polar phospholipid bilayer. The Fluid Mosaic Model In 1972, S.J. Singer and G.L. Nicolson proposed the now widely accepted Fluid Mosaic Model of the structure of cell membranes. The model proposes that integral membrane proteins are embedded in the phospholipid bilayer, as seen in the figure above. Some of these proteins extend all the way through the bilayer, and some only partially across it. These membrane proteins act as transport proteins and receptors proteins. Their model also proposed that the membrane behaves like a fluid, rather than a solid. The proteins and lipids of the membrane move around the membrane, much like buoys in water. Such movement causes a constant change in the “mosaic pattern” of the plasma membrane. Extensions of the Plasma Membrane The plasma membrane may have extensions, such as whip-like flagella or brush-like cilia. In single-celled organisms, like those shown in the figure below, the membrane extensions may help the organisms move. In multicellular organisms, the extensions have other functions. For example, the cilia on human lung cells sweep foreign particles and mucus toward the mouth and nose. Flagella and Cilia. Cilia and flagella are extensions of the plasma membrane of many cells. Difference Between Integral Proteins and Peripheral Proteins Proteins are considered as macro molecules, which consist of one or more polypeptide chains. The polypeptide chains are made up of amino acids bonded together by peptide bonds. The primary structure of a protein can be determined by the amino acid sequence. Certain genes code for many proteins. These genes determine the sequence of amino acid, thereby determining their primary structure. Integral and peripheral proteins are considered as ‘plasma membrane proteins’ due to their occurrence. These proteins are generally responsible for a cell’s ability to interact with the external environment. Integral Protein Integral proteins are mainly found either fully or partially submerged in the phospholipids bilayer of the plasma membrane. These proteins have both polar and non-polar regions on them. Polar heads protrude from the surface of the bilayer while non-polar regions are embedded in it. Usually only the non-polar regions interact with the hydrophobic core of the plasma membrane by making hydrophobic bonds with the fatty acid tails of the phospholipids. The integral proteins that span the entire membrane from the inner surface to the outer surface are called transmembrane proteins. In transmembrane proteins, both ends that project out of the lipid layer are polar or hydrophilic regions. The middle regions are non-polar and have hydrophobic amino acids on their surface. Three types of interactions help to embed these proteins in the lipid bilayer, namely, ionic interactions with the polar heads of phospholipid molecules, hydrophobic interactions with the hydrophobic tails of phospholipid molecules and specific interactions with certain regions of lipids, glycolipids or oligosaccharides. Peripheral Protein Peripheral proteins (extrinsic proteins) are present on the innermost and outermost of phospholipids bilayer. These proteins are loosely bound to the plasma membrane either directly by interactions with polar heads of phospholipids bilayer or indirectly by interactions with integral proteins. These proteins constitute about 20-30 % of total membrane proteins. Most of the peripheral proteins are found on the innermost surface or cytoplasmic surface of the membrane. These proteins remain bounded by either through covalent bonds with fatty chains or through an oligosaccharide to phospholipids. What is the difference between Integral and Peripheral Protein? • Peripheral proteins occur on the surface of plasma membrane whereas integral proteins occur either fully or partially submerged in the lipid layer of plasma membrane. • Peripheral proteins are loosely bound to the lipid bilayer and do not interact with the hydrophobic core in between two layers of phospholipids. In contrast, integral proteins are tightly bound and are directly interacting with the hydrophobic core of the plasma membrane. Due to these reasons, integral protein dissociation is more difficult than peripheral proteins. • Mild treatments can be used to isolate peripheral proteins from the plasma membrane, but for isolation of integral proteins, mild treatments are not enough. To break the hydrophobic bonds, detergents are required. Thus, integral proteins can be isolated from the plasma membrane. • After isolation of these two proteins from the plasma membrane, peripheral proteins can be dissolved in neutral aqueous buffers while integral proteins cannot be dissolved in neutral aqueous buffers or aggregates. • Unlike peripheral proteins, integral proteins are associated with lipid when solubilized. • Examples of peripheral proteins are spectrin of erythrocytes, cytochrome C and ATP-ase of mitochondria and acetylcholinesterase in electroplax membranes. Examples of integral proteins are membrane bounded enzymes, drug and hormone receptors, antigen and rhodopsin. • Integral proteins represent around 70% while peripheral proteins represent the remaining portion of plasma membrane proteins. What Is an Integral Membrane Protein? (with pictures) An integral membrane protein, also known as an IMP, is one which spans the entire biological membrane of a cell. These proteins are attached permanently to the cell membrane, and their function typically relies on being present in the membrane. Both structurally and functionally, they are integral parts of the membranes of cells. Each integral membrane protein molecule has an intricate relationship with the membrane within which it is situated. Structurally, the IMP is usually placed such that protein strands are woven throughout the structure of the cell membrane. Sections of protein protrude through the cell wall inside or outside the cell, or in both directions. The protein molecule cannot function if it is not embedded within the membrane. Another feature of the protein is that these proteins can be removed from the membrane only with very specific chemical treatment. This is because hydrophobic regions of the protein are protected within the phospholipid bilayer of the cell membrane. For this reason, detergents, denaturing solvents, and nonpolar solvents must be used to disrupt the phospholipid bilayer and extract the integral membrane protein. Within the integral membrane protein class are several different categories of protein, many of which are receptors and other types of cell signaling molecules. They are categorized into two groups, based on their structure. These are integral transmembrane proteins, and integral monotopic proteins. Integral transmembrane proteins are those which span the entire cell membrane. These proteins may span the membrane once, or may span it several times, weaving through the phospholipid bilayer such that there are several pieces of the protein protruding through the cell wall. Overall this is the most common type of IMP. Examples of integral transmembrane proteins include voltage-gated ion channels such as those which transport potassium ions in and out of cells. Certain types of T cell receptors, the insulin receptor, and many other receptors and neurotransmitters, are all integral transmembrane proteins. In general, receptors, transmitters, and transporters tend to belong to this class of IMP because proteins that span the entire membrane are typically able to sense conditions both inside and outside of the cell simultaneously. Integral monotopic proteins do not span the entire biological membrane. Instead they are attached to the membrane from only one side, with one end of the protein protruding either inside or outside the cell. This class of proteins includes enzymes such as monoamine oxidase and fatty acid amide hydrolase. Integral monotopic proteins are unable to sense conditions both inside and outside the cell, and are less likely to be involved in intercellular signaling. Membrane Proteins Extraction Membrane proteins (MPs), part of biological membranes, play a crucial role in basic cellular structure and functions, including cell integrity, signal transduction, molecular recognition, material transport, and cell-to-cell communication. Additionally, MPs are the largest category of drug targets. More than 60% currently available therapeutic molecules target one or more MPs. However, there are relatively few MPs with known crystal structures due to the technical challenges associated with membrane protein extraction, solubilisation, and purification. MPs are divided into two major classes: integral membrane proteins (IMPs) and peripheral membrane proteins (PMPs). IMPs are permanently attached to the membrane lipid bilayers. While, PMPs are temporarily associated with either the lipid bilayer or IMPs by means of non-covalent interactions. Generally speaking, IMPs can be purified by more stringent techniques than PMPs, whose extraction just needs a high PH buffer. Here we provide a protocol for IMPs extraction focusing on homogenate preparation, soluble proteins removal, membrane proteins extraction and detergent removal. • PBS: 137 mM NaCl, 2.7 mM KCl, 10 mM Na2HPO4, 2 mM KH2PO4, pH to 7.4. • Homogenate buffer: 250 mM sucrose, 1 mM EDTA, 10 mM Tris-HCl buffer, pH to 7.2, add protease and phosphatase inhibitors cocktail before use. • Blocking buffer: 0.1% (w/v) bovine serum albumin in 50 mM Tris–HCl, 0.15 M NaCl, pH 7.4. • Washing buffer: 50 mM Tris–HCl, 0.15 M NaCl, pH 7.4. • 2% Triton X-100 • Scissors and scalpel blade • Pipettes and pipettors • Homogenizer • Microcentrifuge tube • Sonication • Centrifuge • Columns and appropriate detergent absorption matrix Homogenate preparation 1. For cultured cells: collect cells (0.2-1 x 10 8 ) and wash with ice-cold PBS. Centrifuge at 500 x g for 5 min at 4°C and aspirate the PBS. Resuspend the pellet in 2 mL ice-cold homogenization buffer in a mechanical homogenizer. Homogenize cells on ice. Then skip to Step 3. 2. For tissues: weigh a certain amount of tissues and mince into pieces on ice. Wash with PBS, centrifuge at 500 x g for 5 min at 4°C and discard the wash buffer. Homogenize tissues on ice in 2 volumes of the homogenize buffer, until it is completely lysed. 3. (Optional) Sonicate sample using two 10 second pulses (30 seconds in between pulses) using a probe sonicator. Keep sample in an ice bath and keep probe away from the sample-air interface to minimize foaming. Soluble proteins removal 4. Transfer the homogenate to a fresh 1.5 mL Ep tube and centrifuge at 700 x g for 10 min at 4°C to remove the intact cells, nuclei and cell debris. 5. Collect the supernatant and discard the pellet. 6. Centrifuge the supernatant at 100,000 x g for one hour at 4°C. 7. Carefully aspirate the supernatant (containing cytosol fraction) and collect the pellet. 8. Wash the pellet with homogenization buffer and re-centrifuge at 100,000 x g for one hour at 4°C. Collect the pellet. Membrane proteins extraction by Triton X-100 9. Resuspend the pellet in 1 mL homogenization buffer. 10. Add cells drop-wise to the 2% Triton X-100 while stirring 11. Incubate for 30 min at 4°C with occasional vortexing. 12. Centrifuge at 100,000×g for 30 min at 4°C. 13. Transfer the supernatant to a fresh tube. Note: For peripheral membrane proteins fractionation, you just need resuspend the pellet (obtained in Step 8) in high pH buffer (100 mM Na2CO3, pH 11.3). Incubate for 30 min at 4°C with occasional vortexing and centrifuge at 100,000 x g for one hour at 4°C. Collect the supernatant. Detergent removal by adsorption chromatography Note: Before starting, ensure that your detergent is non-ionic detergent (e.g. Triton X-100) and the molecular weight of protein is large enough to avoid entrapment in the pores of the absorption matrix. 14. Apply distilled water through the column matrix, followed by blocking buffer. Membrane Protein Platform Creative Biolabs has established custom membrane protein and membrane protein antibody production platforms for antibody discovery. Membrane proteins, actually, are a kind of proteins acting as ion channels, receptors and transporters, which enable the cells to transport environmental signals across the biological membranes. Membrane proteins can be classified into two categories, integral (intrinsic) and peripheral (extrinsic) membrane proteins—based on the nature of the membrane-protein interactions (Figure 1). Figure 1. Schematic of membrane proteins in biological membrane. (Molecular cell biology, 4 th edition) Integral membrane proteins, have one or more segments that are embedded in the phospholipid bilayer, of which most span the entire phospholipid bilayer. These transmembrane proteins contain one or more hydrophobic membrane-spanning domains (α helices or multiple β strands), extending into the aqueous environment. Peripheral membrane proteins, do not interact with the hydrophobic core of the phospholipid bilayer. Instead they are commonly bound to the plasma membrane by indirect interactions with integral membrane proteins or by dirct interactions with lipid polar head groups. Peripheral proteins are located to the cytosolic face of the plasma membrane, playing a role in signal transduction. Other peripheral proteins, including certain proteins of the extracellular matrix, are localized to the outer (exoplasmic) surface of the plasma membrane. It has been proven that membrane proteins perform a wide range of functions in cell growth, cell–cell communication (signaling transduction), differentiation, flow of information, metabolism, and migration. Defects in some membrane proteins can lead to diseases, such as cancer, therefore, membrane proteins represent nearly 50% of the targets of therapeutic research. It is very difficult to study membrane proteins due to their naturally low expression levels, although, the study of the structures of membrane proteins is a hot-spot issue all over the world because this will help to better understand the functions of these membrane proteins. Creative Biolabs has many unparalleled expression systems to obtain high-yield and native conformational membrane proteins. Except for membrane protein production, Creative Biolabs can provide customized membrane protein antibody production services. Solubilization of the membrane protein immunogens with detergents may lead to major conformational changes, making their testing with monoclonal antibodies by immune blot and ordinary immunoprecipitation difficult. Using unique MPAT™ and MEAD™ technoligies, Creative Biolabs is specialized in membrane protein and its antibody production services. Please feel free to contact us for a detailed quote. 1. A. M. Seddon, et al. (2004). Membrane proteins, lipids and detergents: not just a soap opera. Biochim. Biophy. Acta., 1666(1-2): 105-117. 2. Membrane protein. ( 3. Membrane proteins. Molecular Cell Biology. 4th edition. Section 3.4. All listed services and products are For Research Use Only. Do Not use in any diagnostic or therapeutic applications. 1.2 Membrane Features The native environment of membrane proteins is dynamic and asymmetric, described by Singer and Nicolson as fluid and mosaic in their 1972 model that became the paradigm for membrane structure (Figure 1.1). 4 Lipids consisting of polar head groups and nonpolar acyl chains form a two-dimensional fluid bilayer (Figure 1.2). Lipid composition is diverse, with most of the lipids randomly distributed in the bulk phase of the bilayer while some are localized by specific interactions with proteins and/or other lipids, often in regions of ordered lipids called rafts (see below). In the lipid disordered phase (Ld, also called Lα), fluidity of the bilayer results from the constant and varied motion of lipids (Figure 1.3). Although the lateral diffusion rate in pure lipid bilayers is very fast, the measured mobility of bulk lipids on the surface of cells is much slower. This difference has been explained by single-particle tracking on the surface of cells containing cytoskeletons: single-molecule trajectories are fast within small regions, where they are confined until they hop to a contiguous region, producing a slower overall progression. 5 The mosaic distribution of membrane proteins results from wide variations in lateral mobility, from those that diffuse rapidly on the surface to those anchored by the cytoskeleton. A large proportion of membrane proteins function in protein assemblies, which themselves have varying lifetimes. Some complexes of membrane proteins are very stable, such as the respiratory complexes involved in energy transduction (see below), while others are the result of transient interactions such as those involved in signal transduction. Many lipid-anchored proteins are observed in membrane rafts enriched in sphingomyelin and cholesterol in the lipid-ordered (Lo) state (see Figure 1.2). The presence of rafts varying in size (diameters from 10 to 200 nm) and in duration (from <1 ms to fairly stable lifetimes) enhances nonrandom distributions and dynamic interactions in the membrane. 6 Lipid–lipid interactions probably drive raft formation, given that even simple lipid mixtures reveal fluid immiscibility (Figure 1.4). The fused hydrocarbon rings of cholesterol are nearly rigid, allowing the sterol to align with lipids containing saturated acyl chains, especially sphingomyelins, and promote the tight packing of the Lo phase. 6 Important Types of Membrane Proteins (With Diagram) Some of the most important types of membrane proteins are as follows: 1. Peripheral (Extrinsic) Proteins 2. Integral (Intrinsic) Proteins 3. Asymmetric Distribution of Membrane Proteins 4. Mobility of Membrane Proteins 5. Enzymatic Properties of Membrane Proteins 6. Isolation and Characterization of Membrane Proteins. 1. Peripheral (Extrinsic) Proteins: Peripheral or extrinsic membrane proteins are gener­ally loosely attached to the membrane and are more readily removed than are the integral proteins. Peripheral proteins are rich in amino acids with hydrophilic side chains that permit interaction with the surrounding water and with the polar surface of the lipid bilayer. Peripheral proteins on the cell’s exterior membrane surface often contain chains of sugars (i.e., they are glycoproteins). 2. Integral (Intrinsic) Proteins: Integral or intrinsic membrane proteins contain both hydrophilic and hydrophobic regions. The hydrophilic portions of the protein interact with the polar heads of the lipid molecules at each surface of the bimolecular leaflet. Portions of integral proteins that project be­yond the surface of the lipid bilayer are also rich in hy­drophilic amino acids. Amino acids in that part of the protein projecting from the outer membrane surface may be linked to chains of sugars. Parts of the protein that are buried in the hydrophobic portion of the lipid bilayer are rich in amino acids with hydrophobic side chains. These side chains are believed to form hydro­phobic bonds with the hydrocarbon tails of the mem­brane phospholipids. It is speculated that within the hydrophobic interior of the membrane, the secondary structure of integral proteins is alpha helix and/or beta sheet (Fig. 15-12). In the alpha helix conforma­tion, amino and carboxyl groups along a stretch of the polypeptide’s backbone form hydrogen bonds with one another in the beta sheet, hydrogen bonds are formed between amino and carboxyl groups in stretches of polypeptide that lie parallel to one another. In the absence of such hydrogen bonding, these amino and carboxyl groups would have polar proper­ties, their hydrophilic nature being incompatible with the membrane’s hydrophobic interior. Alpha amino and carboxyl groups that are not “neutralized” by hy­drogen bonding would be expected only in those parts of the integral protein that extend into the aqueous milieu on either side of the membrane. Integral Proteins That Span the Membrane: M. Bretscher first demonstrated the existence of inte­gral proteins that span the entire membrane. In a se­ries of elegant experiments, Bretscher showed that radioactive ligands specific for membrane proteins of the erythrocyte were bound in smaller quantities to intact cells than to disrupted cells. Disruption of the cells was shown to expose portions of the membrane proteins previously facing the cell interior, thereby al­lowing additional radioactive ligand to associate with the protein. T. L. Steck developed a technique for converting fragments of disrupted erythrocyte membranes into small vesicles that were either “right-side-out” (i.e., the external face of the membrane also formed the ex­ternal face of the vesicle) or “inside-out” (Fig. 15-13). When proteolytic enzymes were added to separate suspensions of each type of vesicle, certain of their membrane proteins were found to be equally suscepti­ble to digestion and could therefore be enzymatically attacked from either membrane surface. These pro­teins clearly spanned the membrane. Other proteins were susceptible to enzymatic digestion only when present in right-side-out or inside-out vesicles, indicat­ing their differential distribution in the membrane’s outer and inner surfaces. Integral proteins that span the entire membrane contain two outer regions that are hydrophilic (i.e., one at each surface of the membrane) the central re­gion is hydrophobic (Fig. 15-12). Carbohydrate asso­ciated with the hydrophilic region facing the cell’s sur­roundings is believed to play a role in maintaining the orientation of the protein within the membrane. The hydrophilic sugars, together with the hydrophilic side chains of amino acids in the out: region of the pro­tein, effectively prevent reorientation of the protein in the direction of the hydrocarbon core of the lipid bi­layer. 3. Asymmetric Distribution of Membrane Proteins: The outer and inner regions of the plasma membrane do not contain either the same types or equal amounts of the various peripheral and integral proteins. For ex­ample, the outer half of the erythrocyte membrane contains far less protein than does the inner, half. In addition, various membrane proteins may be present in significantly different quantities the membranes of some cells contain a hundred times as many molecules of one protein species as another. Moreover, regard­less of absolute quantity, all copies of a given mem­brane protein species have exactly the same orienta­tion in the membrane. The differential distribution of proteins in the various regions of the plasma mem­brane within a single cell was described earlier in con­nection with liver parenchymal cells and intestinal epithelium. This irregular distribution of membrane proteins is known as membrane asymmetry. Not only are the proteins of plasma membranes asymmetri­cally distributed but so too are the proteins of the mem­branes of the endoplasmic reticulum and vesicular or­ganelles (e.g., mitochondria). 4. Mobility of Membrane Proteins: When cells are grown in culture, there is an occasional fusion of one cell with another to form a larger cell. The frequency of cell fusion can be greatly increased by adding Sendai virus to the cell culture. In the pres­ence of this virus, even different strains of cells can be induced to fuse, producing hybrid cells or heterokaryons. D. Frye and M. Edidin utilized this phenom­enon to demonstrate that membrane proteins may not maintain fixed positions in the membrane but may move about laterally through the bilayer. Frye and Edidin induced the fusion of human and mouse cells to form heterokaryons and, using fluorescent antibody labels, followed the distribution of human and mouse membrane proteins in the heterokaryon during the time interval that followed fusion. At the onset of fu­sion, human and mouse membrane proteins were re­spectively restricted to their “halves” of the hybrid cell, but in less than an hour both protein types be­came uniformly distributed through the membrane (Fig. 15-14). The distribution of the membrane pro­teins was not dependent on the availability of ATP and was not prevented by metabolic inhibitors, indicating that lateral movement of proteins in the membrane oc­curred by diffusion. Although some membrane proteins are capable of lateral diffusion, many are not. G. Nicolson and others have obtained evidence suggesting that many integral proteins are restrained within the membrane by a pro­tein network lying just under the membrane’s inner surface (Fig. 15-9). In many cells, this network is as­sociated with a system of cytoplasmic filaments and microtubules that radiate through the cytosol forming a cytoskeleton. 5. Enzymatic Properties of Membrane Proteins: Membrane proteins have been shown to possess enzy­matic activity. Table 15-1 lists some of the enzymes that are now recognized as constituents of the plasma membrane of various cells. To this list of proteins must be added receptor proteins (such as the insulin- binding sites of the liver plasma membrane) and struc­tural or non-enzymatic proteins. Ectoenzymes and Endoenzymes: Enzymes disposed in the plasma membrane may be characterized accord­ing to the membrane face containing the enzymatic activity. Accordingly, ectoenzymes are those enzymes whose catalytic activity is associated with the exterior surface of the plasma membrane the activity of plasma membrane endoenzymes is associated with the interior of the cell. Many (perhaps all) plasma membrane ectoenzymes are glycoproteins. 6. Isolation and Characterization of Membrane Proteins: Because of the relative ease with which they may be purified, the plasma membranes of erythrocytes pro­vided much of the early information on the chemistry of proteins (and lipids) present in membranes. Now, however, plasma membranes can be obtained from many cell types in a reasonably uncontaminated state using various forms of density gradient centrifugation. Nonetheless, the individual protein constituents of the membrane are not so easily extricated for indi­vidual study because of their high degree of insolubil­ity. Varying degrees of success in extracting proteins from the plasma membrane have been achieved using organic detergents (especially sodium dodecyl sulfate, SDS) and concentrated solutions of urea, n-butanol, and ethylene diamine tetraacetic acid (EDTA). These chemicals have a disaggregating effect on membranes and cause the release of many of the membrane pro­teins by dissociating the bonds that link the proteins together or to other membrane constituents. Often, the removal of these agents from a preparation of sol- ubilized membrane proteins is quickly followed by the reassociation or reaggregation of the proteins to form an intractable matrix. Once solubilized, the membrane proteins can be sep­arated into discrete classes using electrophoresis, chromatography, or other procedures. This generally demands that the dissociating agents be present in the separating medium (e.g., the electrophoresis gel, the column eluent, etc.) other­wise, application of the membrane extract to the me­dium is followed by membrane protein reaggregation into insoluble complexes that will not separate into distinct fractions. For example, the separation of liver plasma membrane proteins is achieved only if the electrophoresis gel contains SDS. The solubility problem has been one of the greatest barriers to progress in isolating and fully characterizing the proteins of membranes. Some of the plasma membrane enzymes listed in Table 15-1 have not actually been isolated from the membrane, as removal and isolation of the enzyme is not a prerequisite for establishing its presence. In­stead, the enzyme activity can be measured directly in the (un-solubilized) membrane preparation.
Dog depression after puppies leave (is it real) In this guide, we will explain what dog depression after puppies leave is, and what can be done to help them recover from depression. Dog depression after puppies leave Bonnie Beaver, DVM, Executive Director of the American College of Veterinary Behaviorists, stated vets don’t even understand if dogs experience depression in the similar way we humans do. “It’s impossible to tell and we can’t question them,” added Beaver, who is also a veterinarian expert in animal care at the small animal clinic at Texas A&M University College of Veterinary Medicine. “However there are a few cases in medical care where it might be the only plausible reason.” Dog depression after puppies leave depends on various factors such as age of a puppy, the emotional impact of the dog, and perhaps even her species. It is worth remembering that there is proof that female dogs love their babies. They are capable of understanding and form relationships amongst each puppy. In addition, conception and childbirth improve your dog’s maternal instinct by hormones. This is at its peak during breastfeeding and shortly after birth, but declines with time. That being said, when your puppies are between 7 to 8 weeks old, your female dog can deliberately attempt to escape the company of her puppies. Their teats can continue to get sore due to the sharp bites of the puppy. Its hyperactive behavior would lead her to seek her own refuge. Steady movement from this time will actually provide comfort to your female dog, when she will get more time to herself. She will relax and keep troublesome puppies from running all over them. Even so, removing all the dogs at once will definitely inevitably lead to grieving. Furthermore the split must be phased or the replacement of one litter at a time of re-homing. Yet another research concluded that the more seasoned a mother is, the less emotional connection she feels to her children. In fact, dogs can also experience postpartum depression. Evidence of dog depression after puppies leave Many individuals claim that dogs do not remember like we humans do who our kids or mother is. It is therefore unlikely that the mother dog would experience depression because she misses her puppies. But other studies have shown that the dogs are able to recognize their family members be it their mother or siblings even after prolonged periods of separation. So maybe not in all female dogs but some of them can experience depression after their puppies leave them. There seem to be a number of stories about individuals that tell female dogs recognize their babies and dogs recognize their moms, but then again there are several instances from individuals that suggest the contrary. Galit Reuben, Executive Director of Dogs Without Borders, informed, “I’ve seen mothers connected with babies at adoption functions, and mothers do not always appear to behave in any odd way. When she’s a friendly dog who loves to interact with other dogs, she’ll respond favorably. If she’s nervous and not socialized, she’ll react appropriately.” Although the report indicates that this is not the true essence of the problem. As per Daily Dog Discoveries, scientist Peter G. Hepper performed a research that showed that almost 4 or 5-1/2 weeks of age, the sensory reaction of a dog was good enough to identify its parents and siblings. He further moved on to split the mothers as well as the babies for 2 years to notice how the dogs remembered one another in a prolonged period of time, surly, the mother was able to recognise her already adult dogs, and that they could remember their mom. Decades later, William Carr as well as other Beaver College researchers drew on Hepper’s studies to see whether dogs can remember their parents and siblings after much prolonged periods of absence. The scientists found that puppies who were assigned a blanket that smelled of their mother sniffed at it for a substantially considerable duration compared to a blanket that smelled like that of a stray dog of the very same kind. These reactions still existed in puppies who had not seen their mother in more than a decade. Symptoms of dog depression after puppies leave Dog depression signs are somewhat close to those seen in humans, explained John Ciribassi, DVM, former president of the American Animal Behavior Veterinary Society. “The puppies will be sad. They are likely to be unresponsive. They often start changing their eating and sleeping behaviour patterns. They’re not interested in the activities they once loved.” However doctors caution those signs can however mean that a dog has a health issue, so the primary way to proceed should still be a complete vet check-up. A pet who is sad to be around and doesn’t want to go on a stroll might just have muscle spasms, Beaver claims. Treatment for dog depression after puppies leave Many dogs recover quickly from sadness within several days to a few months with only a little additional TLC, Ciribassi stated. “Keeping them involved, have more of the activities they want to be doing, get them to workout somewhat more, and they’ll be all right,” he stated. Also encourage them when they exhibit signs of excitement, informed Beaver. “When the only activity that always gets some tail wiggle from your dog is a car journey, then bring him on a couple of brief drives a day, thanking and encouraging him when he seems happier,” Beaver suggested. And be mindful not to promote negative actions by providing affection and treatment to a sad dog while mopping, Beaver added. The dog is going to believe you’re thanking them for that action. “Occasionally, if the dog is saddened by the lack of a friend, having another pet will help,” informed Ciribassi. But it has to be handled cautiously, bearing in mind both the interests of the household and the dog, he added. Increase cuddle time Like when we are sad or depressed we enjoy the company of dogs and love cuddling with them. In the same way dogs also love cuddling with us when they feel low. Give attention, love and care for when there is dog depression after puppies leave. Rubbing and scratching behind their ears or tummy is always welcome by dogs. But make sure to not overdo it as dogs like anyone else also desire their time. Play games with them When your dog is feeling sad or depressed, play their favourite games with them. This will help increase their bond with you and make them feel energetic and distracted. You need to remember that being outdoors is always beneficial for the dogs and they look after spending time with you.  Play some music Like us humans dogs too enjoy good music and many researchers have confirmed this theory. Dogs have been seen to respond positively to certain music tunes and rhythms and it does elevate their moods. For dog depression after puppies leave, playing some music may be beneficial especially reggae and soft rock like Bob Marley songs. Medications for dog depression after puppies leave When nothing else helps, antidepressants will assist dogs overcome their depression. Karen Sueda, DVM, diplomat at the American College of Veterinary Behaviorists, suggested stressed dog treatments are just like those often used depressed human beings, Prozac, or Zoloft.Perhaps Clomicalm, an authorised FDA drug can be used for the control of separating distress in dogs. “It’s important that people come to grips with the issue before it gets too serious,” Sueda said. “The moment cases come to me, they’re terrible. But most cases will be successfully handled early on with lifestyle improvements and environmental enrichment, so we don’t have to get to the point that we need to use medications.” Some medications that can help are: Amitriptyline is a medication that is given for the treatment of anxiety and depression as it adjusts the serotonin and norepinephrine chemical levels in the brain and restore the levels to balance. Doxepin is used to treat depression, phobias and OCD behaviour in dogs. Doctors recommend Fluxotine for the treatment of depression, OCD and panic disorders in dogs. Beaver stated it could take up to two months for medications to become successful.But unlike humans who have always been on antidepressants for long periods of time, most dogs will get stronger in six to 12 months and then get rid of medication, she concluded.   In this guide, we explained what dog depression after puppies leave is, and what can be done to help them recover from depression. If you’ve enjoyed the Dog depression after puppies leave mentioned above, I would recommend you to take a look at Dog depressed wearing cone and Husky Depression too. What we recommend for Depression Professional counselling FAQs: Dog depression after puppies leave Do dogs get depressed when their puppies leave? Yes, dogs get depressed when their puppies leave. They’re already going to be unhappy to have “lost” their mother; thus, splitting pups from one another is unfair, and that’s what makes them sob at home for the first several days. Dogs don’t have the same social relationship as people have with their relatives. A dog will be totally comfortable to be bred away from his litters. Do Puppies feel sad when they leave their mom? Frequent crying, howling, screaming, or finding a puppy continuously underfoot might be an effort, but please remember that your new pup has since been removed from the comfort and protection of everything he’s ever experienced. Being anxious after losing his mother is absolutely normal. Why is my dog leaving her puppies? If your pet has been sleeping or staying away from the litter for a long time, it is possible that she might have refused it. She will also exhibit symptoms of tension, and in some situations she may literally pick up and carry the babies away from her. An even more strong indication of possible dismissal is that whenever you hear the litter’s prolonged crying. Do dogs miss their moms? Some claim the dogs can’t recall their mothers. He will look at the possibility that the male dog is going to partner with the mother as confirmation that the dogs have no awareness of each other. The puppy, though, has no conscience. If the puppies were put in the area, 84% of the time they reached their mothers. Can a puppy stay with its mother forever? Puppies do not really remain with their mothers always. If a puppy is split too soon, there could be complications, like attachment disorder. Even though it’s not adopted too early, but is placed apart from all other animals, precious social interaction time may be wasted. Do dogs miss their littermates? Although some scientists think that a dog may not remember their litters, there is also some proof to demonstrate that this is not accurate and that they can possibly do so. Steven R. Lindsy’s hypothesized dogs have the capacity to recognize their offspring later on in life and can sometimes recognize their family. Was this post helpful? [Sassy_Social_Share type="standard"]
A valve that allows water to stream in one direction and will then close to prevent development of a back-flow. Check valves function on flow sensitivity in order to propel opening or closing. They are generally made up of two discs, one which allows for a forward flow and the other not allowing backward flow by shutting the valve when there is a reduction or fluctuation in the flow rate. One of the main determinants of check valves is that they are unable to be utilized in pulsating systems due to the valve being insufficient in preventing forward flow operation as a result of the vibrations.
Is Computer Science a Language? Lauren Salzman  When you think of choosing a language to study, what’s the first thing that comes to mind? Ruby? JavaScript? Python? Probably not, because all of the aforementioned languages are programming languages. Across the country, there is a debate escalating about whether computer science should be counted as a language credit. Should we be able to take computer science instead of a traditional language like French or Spanish? This debate has been popping up in schools everywhere, and even in state legislature. “There are a lot of interesting similarities between learning a natural language and a programming language,” explains Ms. Hansberry, one of Latin’s computer science teachers. Both genres of languages require the student to learn new vocabulary, both have grammar and syntax, and both require knowing the rules behind the language. In our world with technology at the forefront, having students take computer science can better help them prepare for technology-related jobs after college. Even non-STEM jobs are starting to acquire more STEM employees, due to the many applications of math, engineering, and computer science. Ms. Hansberry comments that “[she has] also heard the argument that knowing how to program is just as useful (or even more useful, depending on who you ask) as knowing another language, so students should be given the option to learn this… skill set instead.” This also offers a way for students who have speech disabilities or impediments to still be involved in a language program. It is undeniable that millennials interact with technology and code more than their parents. However, if this debate is taken to state courts, their parents are the ones who are making the decisions. The generational need for languages is evolving. While generation X, the post baby boomer generation, never thought of computer science surpassing natural language in importance, millennials have a different perspective on what it means to communicate through language and technology. But is making computer science count as a language credit the best way to do this? Ms. Hansberry remarks, “I ultimately, however, don’t think it’s a good idea to offer computer science as a language. I have a bunch of reasons why. The first is that computer science is a lot more than just learning a programming language. Depending on the topic, studying computer science has just as much in common with studying math or chemistry as it does language. In addition to learning programming languages, in computer science you also learn how computers and their software work, how to encode data in a digital format, how to design and analyze algorithms for solving problems, different tools for writing software, how the internet and networking works, etc.” At Latin, students already have busy schedules with mandatory requirements, along with a seemingly endless list of elective they wish to take but have trouble fitting in. In an ideal world, computer science would be a mandatory class, but placing another required class would lead to more stress, strain, and less choice in one’s schedule. “For schools that want to find a way to encourage and allow more students to take computer science classes, allowing them to take it instead of a language will free up time in their schedule to do so,” says Ms. Hansberry. This is the main reason that some argue computer science should be fulfill a language credit – so it is more accessible. But Ms. Hansberry ultimately explains that computer science is not a natural language, but rather an invented one, and should not be an option for fulfilling a language requirement. “There is a lot of value in learning a natural language,” she says, “Human brains learn natural languages in a different way than they learn invented languages. People who study the brain and language think that learning a natural language is a unique process.” For example, an invented language, called Esperanto is the most widely spoken constructed language in the world. Some children have adopted it as their second language, but studies have shown that children learn this language differently than they would a natural language. Although the reasons are not known, the evolution of language is natural to humans and has a cultural component that is not available with constructed languages, such as Esperanto and those you learn in computer programming. The struggle with having computer science as an elective, is that many students are not able to fit it into their schedules and thus do not realize and reap the benefits of learning to code. With new programs such as Girls Who Code, and Codecademy, and the increase in jobs that require STEM professionals, new generations will most likely grow up with coding as a significant part of their education. In fact, the Lower School computer science program has introduced coding as part of its curriculum. But for current high schoolers, who are growing up as coding is starting to take center stage, it is challenging to find a place for it to fit. So instead of taking computer science in lieu of Chinese or Latin, the goal is to find time in your schedule to keep up with this evolving language – whether that be teaching yourself online, going to a computer science summer program, or signing up to take a class with Ms. Hansberry.]]>
teckies.com files h t t p : / / w w w . t e c k i e s . c o m / f i l e s Search the Internet by Khalil Yazbeck August 03, 1999 A Search Engine is a tool or a software that helps a web surfer locate specific information on the web. Normally, a search engine has a search box that looks like this: type your text here then press search In addition to a list of general subjects to restrict your search. e.g Universities, Law, Medicine, Computer, Internet, Health, Games, etc.. The most famous search engines on the Internet are: Yahoo, HotBot, Altavista, Lycos, Webcrawler, Infoseek and Excite. Search engines are grouped into 2 categories: • Directory search engines. • Index or Spider search engines. Directory search engines are used when searching for general information. Usually, the search results are displayed and grouped into categories.  If you are searching for Universities, Companies or Countries use directory search engines. Yahoo is the most popular directory. Index or Spider search engines are used when searching for something specific. The search engine searches the content of a webpage stored in its database, so when querying a search engine like Altavista you're actually searching its database and not the whole Internet. The search engine might return website addresses or just individual pages of a website matching your query. If you're looking for "food" you'd better use a Directory like Yahoo, you will get irrelevant results when using index search engines because the index search engine matches the words you have entered in the search box with the content of a website, many irrelevant websites might contain the word "food" and appear in the search results page. To find your way through more than 300 million webpages on the web and get what you want, you should master some basic searching techniques. The following are searching techniques applicable to almost all search engines: - Phrase Searching:  Use it when searching for a specific subject. For instance, if you're looking for Internet Solutions you should type "Internet solutions", put the 2 words between quotations, so you will only get webpages containing both words (next of each other). Suppose you've typed internet solutions without the quotes, what you will get is pages that contain the word internet and other pages with the word solutions, most probably the result page will contain thousands of irrelevant pages. - title searching: Searching by title is not supported by all search engines, however it is very useful with Altavista and allows you to get accurate results. By typing title: followed by the search word or phrase, the search engine will only look for pages that contain the word or phrase in the page title (this is the title you get at the top of a webpage e.g teckies.com online magazine). The logic behind this technique is that websites that talk about the dangers of smoking should at least contain the word smoking at the title. - Boolean Operators: Maybe you wonder what math got to do with search engines, well boolean operators come very handy when searching on the Internet. By combining operators such as AND, OR, NOT you can control the results you get thus search more efficiently. Let us take a look at those magical operators. • AND: Use the Boolean AND in capital letters between your keywords and you will only get webpages that contains both words. If you want to search for apple pie just type apple AND pie. • OR: OR operator is used between keywords to get pages that contains either of the keywords. For instance typing bike OR car will get webpages that contain the keyword bike and others that contain the word car. • NOT: NOT operator is used to eliminate pages that contain a keyword. For instance typing Clinton NOT Monica will dump all pages that mention the Monica Lewinski scandal. In addition to Boolean operators you can use (+) to add keywords just like AND and minus (-) to discard certain keywords from search results. Make sure you don't use any space between the sign and the keyword e.g +space -nasa will get webpages that contain the word space and dump those with the word nasa. P.S: All Boolean operators should be capitalized, otherwise they will be treated as a keyword. Which technique gives better results ? There's so magical formula to get what you want from a search engine, it depends on what you're looking for, you are often advised to try several techniques to see which one yields better results on a particular search engine even combine techniques. For instance your can use phrase searching with boolean logic e.g "space travel" NOT "fiction" this will return pages that only contain "space travel" phrase and dump all pages with the word "fiction". As mentioned earlier, title searching is very efficient especially with Altavista, for instance if you want to look for some Windows tutorials on Altavista, simple type title: windows tutorials and see what you get. The best thing to do when searching on the Internet is that you should stick with one search engine, don't spend time jumping from one engine to another, instead take a look at the search engine's help page to see which searching technique is supported, sometimes the help page can give you a hand. Try to look for specialized search engines, although Altavista is one great search engine sometimes you could find what you're looking for on a local search engine. I usually start my search by looking for a specialized search engine and I often find what I want. About teckies.com | Advertise | System Requirements
The Beginners Guide to Understanding Your Furnace Back to All Articles As we head into fall and winter, it’s a great idea to learn about the machine that provides heat and comfort throughout the cold winter days – your furnace! The more you know how a furnace works, the better you can troubleshoot it if something goes wrong. Energy Costs A rise in energy costs is not just down to external factors, it can be down to the furnace losing its efficiency and wasting more energy than it is using. A repair can do the trick, but if repair prices are over one-third of the furnace’s worth, then you should seriously consider replacing it. Another indication that your furnace needs replacing is if it’s making unusual noises such as rattling, popping, screeching, or humming. It also depends on which part of the furnace these noises are coming from. The Thermostat Your furnace starts when it receives a signal from the thermostat that tells it to turn on. Depending on the temperature you set, when the thermostat detects the air temperature dropping below that number, it activates the furnace. When the thermostat sends its signal to the furnace, the furnace gas valve opens and ignites the burner component beneath the combustion chamber. The gas valve works with the thermostat to regulate the amount of gas that flows into your furnace. Furnace Maintenance Tips Schedule a professional tune-up every fall. Your HVAC specialist will be able to test your heat exchanger for cracks, clean your pilot light, inspect the ventilation system, and much more. Replace the furnace filter at least once every 90 days (set reminders to check the filter every month). Write the date of replacement on the filter itself so you know exactly how long it has been since the last filter change-out. Each furnace is individual. If you have any questions or concerns regarding your current furnace, feel free to get in touch with a Trane representative. How Indoor Air Quality Can Affect your Child’s Health How Indoor Air Quality Can Affect your Child’s Health Previous Article
Learn From History A nation must learn from history valuable lessons which nobody can teach. History is not an account of kings or rulers but political trends  and events causing social upheaval and economic development or decline. India has limited study of history to patches constructed in the most partisan manner. It is fake, unless complete account of both happy and unhappy events and trends is included in a fair and just manner with no bias towards anybody. India must promote a comprehensive study of history, with emphasis on current history covering the last 200 years. Special focus has to be on history of post-independence period, pre-independence period divided in sections on every decade of the twentieth century culminating in the tragedy of partition in 1947. If done honestly, it will throw up results sanitising history of misconstruction of developments of the period, which unfortunately has been sold by irresponsible historians as history, though none of them really was qualified as a “historian” in the true sense of the idea. Any study of Indian history would remain inconclusive without simultaneously studying geography and historical-geography, anthropology, cultural islands self-reliant and complementary to neighbouring islands making collectively a nation called Bharatvarsh. Then comes another significant part of history: the world history. Study of European history is essential. So is the study of rest of the world between Europe and Bharatvarsh. Finally, America, Canada and Russia need to be studied. Currently, our history-challenged politicians keep creating unnecessary chaos by picking up some word like Fascist or Nazi without understanding the idea of fascism and its history from birth to growth and decline or Nazism similarly. That explains the lack of understanding the developments in Afghanistan, the history of turbulence there, the influence or pressures from neighbours and the region, the influence of developments globally, the source of strength the resistance of the Taliban draws upon, the games super-powers have been playing in the country or the impact of fast changing geo-political scenario affecting developing countries since WW-I and particularly the WW-II, the Cold War and post-cold war  situation. Similarly, the reasons why and how the Taliban using technology have succeeded in depriving the masses of such instruments of empowerment and growth. What do the developments in Afghanistan mean for the world? The USA created the situation by holding direct talks with the Taliban over the head of the legitimate government and compounding its mistake by making the statement that the Afghani people have to look after themselves. Who does not know the people of a country have to look after themselves? Do the people of other countries routinely look to America for help or does the USA come forward voluntarily to the help of suffering people in other countries? The reality is that it interferes in other countries, fuelling instability and turbulence, only to fish in the troubled waters. If it does, it pursues its own interests only and no charity to other countries. America should not have entered Afghanistan in the first place and if it had landed there it should have got out intelligently. It has created only a serious threat to world peace. It has accepted terrorism as a legitimate instrument of state policy. And that is horrible. It will not shield America. Rather, the threat to America will grow. Terrorists don’t take other countries seriously. Their target has always been America. And they have won the battle and America has been defeated woefully. Pakistan stands to gain enormously. It shall continue to be the world’ terrorist factory. It has so far benefited from engineering insecurity in democratic nations. Now it has the affection and care of China to support its policies and games. China will try to derive benefit from the defeat of America in Afghanistan. It will have a free hand in Balochistan and Afghanistan, where it will capture territories and rich minerals. It will pose a threat to Russia by containing Russian influence in the region. Wide spread regional and global changes are bound to occur as a consequence. One, it is a proof that the idea of the United Nations is practically dead now. Matters stand decided through armed conflicts and world history will undergo change. Two, the international trade will suffer and shall need to be reset to meet the collective challenge of Afghanistan, Pakistan, China and terrorist organisations like the ISIS, Al Qaeda etc. Three, and this is the most important, the world will be compelled to acknowledge the real source of the threat to world peace: Islam, Shariya and its interpretation or misinterpretation by the Mullahs. Four, chances of intra-Muslim countries conflicts are bound to grow and come in the open. Iran can’t be isolated nor can the Shia sections of Muslims be subjected to violence by the Sunni sections, nor women can be subjected any longer savagely in the Taliban style. Five, the enlightened sections of the Muslims world over will have to come forward and speak out their minds openly against the barbarity in the name of Islam. They are a very big and powerful population of more than 50 million men and women, who are highly educated, professionals from all field of knowledge and rich enough to stand up to any organised violence against the helpless women and children. The next decade will be a test of the strength of democracy and all the values vouched so far for a healthy democracy in the interests of humanity. Terrorists shook America on 9/11 as the most powerful nation was caught unawares by a clutch of terrorists. The myth of invincibility of the world’s top military power was demolished in a few minutes as the news of the damage started coming. But it was not the death of thousands of innocents or demolition of the trade towers alone but it ruined the American economy, imposed a very heavy cost on homeland security, pushed America to a meaningless war in Iraq and Afghanistan. Pakistan made a fool of America, robbed billions of Dollars in ransom promising to help America catch Osama Bin Laden while keeping him in safe custody in its Army Cantonment in Abbottabad. The events since 9/11 affected most countries as the world economy, trade and security suffered. The Afghanistan developments will affect America and the rest of the world even worse for decades. All the values of democracy will get redefined before terrorists trample upon nations in Europe, America, Asia, Africa, Australia, New Zealand. With help from China, they will come to possess nuclear and biological weapons. That is a real threat and no imaginary conclusion. The good realize the evil acts only after a 9/11 happens! 1. Leave a comment Leave a Reply WordPress.com Logo Google photo Twitter picture Facebook photo Connecting to %s %d bloggers like this:
Dark Origins – A Christmas Carol by Charles Dickens In the spring of 1843, Charles Dickens read a government report on child labour in the United Kingdom. The report, compiled by a journalist friend of Charles Dickens, comprised of a series of interviews with working children. It detailed the long hours, crushing work, and poor conditions suffered by these children. The new and heartless attitude towards child labour was a result of three things: • an increase in the population by 64% in 30 years; • workers leaving the countryside and crowding to the cities in search of work; and • the demise of cottage industries and there replacement with mundane and menial labour in factories. Employers thought of the workers as commodities whose labour was measured purely on output and productivity. There was a lot of controversy among the wealthy classes and the clergy as to whether assistance should be extended to the poor. A lot of people were of the opinion that people were poor due to their own laziness and malingering and that giving help would exacerbate these tendencies. The work houses of the day split up families, provided minimal food, and extracted hard labour from its occupants, including children, in an effort to discourage the poor from seeking help. I am reminded at this point of the song Food, Glorious Food from the musical Oliver based on the book by Charles Dickens: Rev. Thomas Malthus advocated letting the poor go hungry to decrease the population. His view was that it was better to let the poor starve to “decrease the surplus population”. Charles Dickens’ response was to write the novella, A Christmas Carol, which eloquently expressed his views on employer responsibilities towards workers. If you don’t know the story of A Christmas Carol, this is a very brief overview: The story opens with Ebenezer Scrooge sitting in his counting house on Christmas Eve. His clerk, Bob Cratchit, is sitting shivering in the anteroom because Scrooge won’t spend any money on heating. He turns down his nephew, Fred’s, invitation to a Christmas party and chases away two men collecting money for charity. At the end of the day, he returns to his cold, dark home. After Scrooge has retired for the night, he is visited by the ghost of his dead partner, Jacob Marley. Marley is weighed down by heavy chains and is destined to make his way through the afterlife dragging them after him because of his mean-spirited and selfish life. Picture caption: https://www.charlesdickenspage.com/illustrations-carol.html Marley tells Scrooge that he will be visited by three ghosts that night, the ghosts of Christmas Past, Present, and Future. The ghosts show Scrooge where he made mistakes in his past life due to choosing money over love and life, how his clerk and the Cratchit family are suffering because of his present day meanness, and show him a lonely future death. Scrooge is offered, and takes, an opportunity to change his ways and find redemption. If you are interested in listening to A Christmas Carol beautifully read by Stephen Humphreys, you will find the links on Rebecca Budd’s blog: Clanmother: https://clanmother.com/2021/12/07/stephen-humphreys-reads-a-christmas-carol/ Wishing you all a Merry Christmas if you celebrate or Happy Holidays. Although I cannot compare my take on Victorian child labour to Charles Dickens’ brilliant works, I have written several times about this and I thought I would share this short extract from my book, Through the Nethergate, about a serving girl in a tavern in Bungay in 1589. “The rich, amber fluid flowed into the waiting tankard, in striking contrast to the damp, darkness of the barrel filled cellar. The small, frail girl stood with the tankard in her trembling hand. She was hungry, thirsty and cold. She hadn’t had anything to eat or drink since last night’s frugal supper of leftovers in the Inn’s kitchen. A wave of dizziness washed over her as she contemplated the drink. Its golden depths seemed to entrance her as she lifted it to her lips. At least the kitchen was warm, she thought, remembering the delicious heat of the enormous, roaring fireplace. The kitchen was a much better place to steal a moment of rest than this freezing cold cellar, in the bowels of the building. The strong, rich taste of the ale brought a smile to the girl’s pale face. She greedily drained the tankard, closing her eyes and allowing a feeling of well-being to permeate through her skinny, undernourished body. The girl, called Lizzie, worked as a servant at the pub and she was twelve years old. She knew she should be grateful for the job, but it was hard to forgive the heavy-handed punishments metered out to her by Will, the owner of the establishment. A rough hand grasped her shoulder, its thick fingers digging viciously into her flesh. “What have you done?” the loud, grating voice of Will blasted through her euphoria. Lizzie jerked with fear and the tankard fell from her fingers, clattering to the stone floor before rolling away. She looked up into piggy eyes staring out of a fat and well-fed face. Will’s usually florid complexion looked even ruddier and coarser than usual. “Why, you little thief,” continued Will. “You know what we do with thieves in this Inn.” A short while later, Lizzie found herself chained to the wall of the cellar. Her pleas and cries for mercy had fallen on deaf ears as Will, filled with righteousness and piety at her ungodly action, attached the manacles to her wrists and ankles.”
B4A Library ICUB4A - detecting character-encoding formats When experimenting with subtitles, I noted that very often characters were not displayed correctly due to wrong character-encoding being used during the loading of a subtitle file to display. You have surely all seen and heard of UTF-8, ISO-8859-1 and so forth. These abbreviations represent various character-encoding formats. Unfortunately, there is no 100% method for a program/application to know which character-encoding format to use when loading a file unless the file-format provides it. Therefore, most software uses a detection-algorithm to guess the character-encoding format in use. I tried several detection-algorithms but at the end, I decided to use ICU4J which is continously updated and maintained and which according to my tests furnished the best results/guesses. The size of the original ICU4J-library is very large (it provides other functionaility as well) so I made a wrapper for B4A and B4J using a subset of the APIs available (total size is approximately 75KB). Anyway - what is ICU4J? ICU4J is an open-source, widely used set of Java libraries providing Unicode and globalization support for software applications. Companies that use ICU4J are for instance Google, Apple, Adobe, Amazon (Kindle), Debian and so forth. Google has announced that with Android N they will also include a subset of ICU4J's APIs. Usage in B4A: Sub Process_Globals 'These variables can be accessed from all modules. End Sub Sub Globals 'These variables can only be accessed from this module. Private guessEncoding As IcuB4A End Sub Sub Activity_Create(FirstTime As Boolean) Dim fileName As String Dim fileLocation As String Dim detectionResult As String fileName = "bbchinese.srt" fileLocation = File.Combine(File.DirRootExternal,fileName) detectionResult = guessEncoding.readFileAsStringGuessEncoding(fileLocation) End Try End Sub I'm attaching sample-project, test-files and library. You can use this library with any text-files. I have also posted a wrapper for B4J which you can find here. I hope it may be useful for someone. • ICUB4Asample.zip 6.8 KB · Views: 135 • libs.zip 62.9 KB · Views: 183 • testfiles.zip 45.5 KB · Views: 113 Last edited: Licensed User I'm using the sample but I put the bbchinese.srt into my dir asset. fileLocation = File.Combine(File.DirAssets,fileName) I got this result: (ErrnoException) android.system.ErrnoException: open failed: ENOENT (No such file or directory) Licensed User The error you mentioned is likely caused by the fact that the Assets directory is read-only. You will need to copy the file to somewhere you have write access and then call the detection. You could create a sub like this (not tested) if you know that the file is in the Assets directory: Sub another Dim detectionResult As String = GetDetectionResult(fileName) end sub Sub GetDetectionResult(Filename As String) As String Dim filelist As List Dim i As Int For i=0 To filelist.Size-1 If Filename = filelist.Get(i) Then Dim fileLocation As String = File.Combine(File.DirDefaultExternal,Filename) Dim detectionResult As String = guessEncoding.readFileAsStringGuessEncoding(fileLocation) 'delete the file after checking If File.Exists(File.DirDefaultExternal,Filename) Then End If Return detectionResult End If End Sub
School News Digital books increase access to the power of reading Photo by Teens in Print Words are such an interesting thing. We use them to communicate what’s on our minds, and essentially, they’re how others see and remember us. Reading, however, wasn’t so easy for me. Some types of texts were boring to read. Some books had words I’d never even heard of before and they had so many pages that I would never be able to get through them. As a kid, I tried to pick up reading because I wasn’t good at English and it was a way for me to build up my skills. Reading was super difficult. I used to wonder why some of my friends liked to read when I always thought it was so boring.  Recently, I’ve tried to pick up more books to read for pleasure and I like the experience a lot more now. It’s almost something I picked up accidentally. My friend and I talked about the app Libby one morning before class started and I began to think, “Well, one of my goals for 2020 was to get into reading.” With Libby, you can link the app with your library card and there’s a bunch of e-books you can read. I knew that I was stressed at school and I had a bunch of free time on public transportation, so I could try picking up a book and get into the habit of reading.  According to the American Psychological Association, “less than 20 percent of U.S. teens report reading a book, magazine, or newspaper daily for pleasure.” It’s not a huge shock, seeing that as teens, we’re usually always on the move from school, extracurriculars, work, whatever it may be, so it’s rare to have tons of free time. What I found when I started picking up reading again is that it actually reduces my stress and has acted as a way for me to stay sharp. I’ll use it as a break when I’m doing homework, it keeps me occupied when I’m taking the bus home, and honestly, it just gives me something to keep up with. Words are timeless, and it’s important to read because in a time with so much political chaos and mass media, we must remember how to think, analyze and act for ourselves.  In Alisha Rajpal’s TED Talk, “Read More. Think More,” she discusses how reading has been impactful on her life as a youth in today’s world. She says to the crowd, “Consider you’re reading a book and there’s a line that goes ‘an attractive man walks down the street’ ... for each of us that attractive man could look different.” When we read, we have our own interpretation of the world as compared to seeing it on screen where society’s norm of an attractive man would be pressed upon us and accepted. When we read, we use our own imagination, our own beliefs and, honestly, our own brain.  The app Libby is amazing. I found that my main problem with reading was that it was so hard for me to constantly pull a book out and put it away if I wanted to read for only a short amount of time. I’m on my phone all the time, and Libby has made it convenient and easy for me to search for a variety of books to read instead of doing things like scrolling through Instagram for hours. On Libby, I just finished reading “David and Goliath” by Malcom Gladwell and now I’m ending February with “The Goldfinch” by Donna Tartt.  So here’s my message to you: pick up a book! Explore what interests you, go through adventures and misfortunes with fictional characters and don’t fall into the societal norm of not reading. Reading lets us build our own character in this world.  Read more… School News Finding studying difficult? Use these tips! Teens in Print Do you enjoy the spring? It’s a time in the year when new life begins, where people can freely go outside and enjoy the beautiful, blooming flowers. But for students? Nope! We don’t have a chance to do all that. Instead, we are stuck with exams such as the MCAS, SAT and finals. The best way for us to do well on these tests is to review our notes while slouching our backs at a desk, also known as studying. Standards-based assessments, such as the MCAS and the SAT, require lots of studying. Taking these tests can be challenging for many students, especially with the fear that if they don’t know the material, they could potentially hurt their grade or struggle to succeed academically. Moreover, when students stress out too much about their studies, they could perform badly as a result or put their mental health at risk. In this article, we will be sharing tips students can use while preparing themselves for exams. This way, we can all do our best without getting too stressed.  Tip 1: Don’t Only Study At A Desk, Study Actively!  Hunching your back at your desk is tiring, I know. The best posture to have is sitting up straight, however, having that posture makes it difficult to reach your desk. According to Psychology Today, constant sitting puts you at risk for heart disease, diabetes, depression, obesity and other medical concerns. To avoid this, be active while studying instead of only sitting! Put your body in an active position while studying: CBS News reports that, “Standing while learning and completing assignments improves executive functioning, or the skills you use to break down tasks like writing an essay or solving multi-step math problems.” This proves that keeping your body in an active state helps us study better, because studying requires a high demand for your brain’s functions. Moving will keep your brain working, making your studies effective. Exercise! Have a stretch!: CBS News also found that regular exercise has been shown to boost verbal memory, thinking and learning. It improves alertness, attention and motivation, while helping to build new brain cells. Make it fun!: Instead of using piles of class notes to review, add a little twist to it. You could turn your studies into a sort of game, like a “Family Feud”-based vocabulary game, or a board game or whatever else you can think of! As a bonus, invite classmates to play your game, so you’ll not only be having fun, but studying as well! Tip 2: Be Organized! If you’re often disorganized, you’ve probably had the experience of being in a rush, cramming things last minute or pushing yourself to the limit while multitasking. According to Productivity 501, a website dedicated to offering productivity tips, good organizational skills lead to productivity and efficiency. Appearing organized is one thing, but actually being organized when the work gets tough will make all the difference. Here are some organization skills that should be used for studying: Having a clean environment: Working in an environment that is neat and tidy will reduce the amount of time you spend hunting files down, and increase your ability to get things done.  Get rid of and avoid all distractions: Technology, your friends and even the environment where you study can all be distractions. Yes, studying while listening to music can help, but let’s be real, we end up doing other things on our technology instead of studying. To avoid these distractions, isolate yourself. Leave your mobile phone, tablet, PC and everything else that could distract you out of the room. Find a quiet room that lacks any environmental noise that irritates you. Make sure that room doesn’t have anything that will detract from your studying. If it is a mess, you might end up cleaning it all up. If there are many visual distractions, you might find yourself staring off into space at something in the room. When you truly find the right space for you, there won’t be anything to waste time on and you will progress much faster in your work. Planning out your daily schedule: Schedules are especially important for students with extracurricular activities like sports, after school clubs or jobs. Oftentimes, these students are limited in their time to study. According to the University of Michigan at Flint, the expected study time for high school students on average is around one to two hours per night. One to two hours isn’t that much, so you’ll have enough personal time to add to your schedule. Try to make sure your study time is when you feel most comfortable, or in the mood. Tip 3: Eat Brainy Foods! To those that like to eat, you’re in luck. It’s good to eat healthy while studying to give your brain a boost (but don’t leave a mess on your paper!) According to, these foods listed below are brainy foods you might like: Fish: Salmon, sardines and mackerel have lots of protein and Omega-3 fatty acids, which can help your brain’s memory functions. Eggs: Here you can find important proteins and vitamins like B6, B12 and choline, which help regulate mood and improve mental function. Green Tea: Tea is a good alternative to coffee or sodas full of caffeine, which students normally drink to keep themselves awake while studying. This beverage provides antioxidants that help sharpen your concentration. Fruit: Sugar is sweet and addicting, I know. But fruit is sweet and addicting and beneficial, too! It’s a healthier alternative to a candy bar and provides nutritious vitamins for your studying brain. Tip 4: Support Your Quality Of Life!  You want to know what the most important part of studying is? You. You are the one that’s going to take this exam. Therefore you are the one that’s going to study all of these topics. And for this, you’re going to need to put yourself in the best condition! Take a look at these tips to help you out: Get comfortable: According to Ameritech, a healthcare college, “Numerous studies have found that the place—or ‘context’—in which you study affects how you’ll remember that information.” The better-suited place you study in, the better you’ll remember information, while the least-suited place can lead to the opposite. Make sure your study environment is the most suited for you! Make sure it has things like comfortable lighting, or nice stationery or whatever you like. Another way to get comfortable is to take breaks every so often. If you take frequent breaks every interval of study, then you will stay refreshed and won’t become mentally exhausted.  Combine your studies with your creativity/hobbies: You’ve heard the phrase “unleash your creativity” right? Well, if you want studying to feel easier, try to connect it with a hobby. Since you’re already used to your hobbies, combining them with your studies can help make learning both comprehensive and entertaining. For instance, if you like to sing, then why not compose a song based on your study notes? If you like to write, why not write a story based on your notes? If you like to act, why not make a story based on your notes, and act it out? There are many possibilities, so try it out! Study in a good mood: As I implied, no one likes to study. But even so, don’t force yourself to study while you’re in a bad mood! Make sure you’re not overwhelmed with anything at the moment, and you are feeling completely energized and organized. Maybe even prepare a reward for yourself so you can be more motivated to study! Psychology Today found that a positive attitude has a specific influence on learning. A good mood affects your ability to learn, especially subjects that require some amount of flexibility and creativity. So, next time you are hitting the books, try to remember some of these helpful tips. You might be surprised by how much you improve and succeed! Read more… School News Consider community service outside of the animal shelter There are other organizations with more flexible volunteer schedules that need your help Community service can be daunting. Similar to the college process, community service is often put off until we get closer to graduation. Counselors see students as invisible until senior year, and then pounce on them to demand college applications and graduation requirements. You may be asking, what is community service? Community service requirements vary from school to school. My school, Snowden International, requires students to perform 100 to 150 hours in order to graduate. Community service is when an individual offers a service to the neighborhood or certain nonprofit organizations, like local animal shelters or homeless shelters. Community service is not paid — any form of monetary compensation means it's not community service. But how do we find a location where we can serve?  High school youth often have a hard time finding community service opportunities. Competition with college students, age restrictions and parental requirements all make community service harder to find than it should be. Animal shelters are the usual go-to for high school students. Shelters like ARL Boston and MSPCA-Angell take volunteers seasonally. Unfortunately, these positions tend to have a screening process and a tedious application process that can make it difficult for students. The competition with veterinary-students-in-the-making makes it nearly impossible to snag a spot. However, there are plenty of places to volunteer around Boston!  First, did you know that you can get community service hours from your job? Oftentimes, unpaid work or community-oriented jobs will count as service hours. These opportunities don’t come along all the time, so it’s always a good idea to ask your employer. Massachusetts General Hospital, for example, offers community service hours for teens in Boston on an application basis. The application opens in the fall of each academic year. Also, local shelters will always take in volunteers! Rosie’s Place, a women’s shelter in South End, accepts volunteers year-round. You can find flexible hours and self-schedule as an individual volunteer, and often you get free snacks too! Serving as an individual volunteer builds confidence — you can’t rely on your friends to pull the weight for you. On their website, there is a short application. After it is reviewed, you’ll be given an account to schedule yourself as a volunteer. Most volunteer options are doable as a teen, and there is always a supervisor available to assist with questions and teach you new tasks. At Rosie’s Place, there are two main options for youth volunteers in the kitchen: preparing and serving food or assisting in the food pantry. Youth can also help sign in visitors. In the kitchen, I alternated between washing dishes in the back and serving and cleaning up in the front.  Surprisingly, community service is broad enough to find an opportunity in the area you’re interested in. Community service gives a sense of responsibility to the volunteer. Having people dependent on your work makes you appreciate the features in our communities that we usually overlook. Animal shelters are super-competitive to even get an interview, so if you’re in a time crunch, applying to a local shelter is your best bet. Local shelters offer less-rigid hours, and more individuals are able to register as volunteers.  Read more… City News Advice on handling minimum wage jobs If you attend a public school in Boston, it is likely that either you or someone you know works at some form of internship or minimum-wage job. For many students and adults, these jobs are a convenient way to make money, but working in a position viewed as disposable by a large franchise can have a number of downsides. This can be especially problematic for workers who rely on the income these jobs provide and don’t have the option of leaving if it becomes a toxic work situation.  Many examples of this abusive behavior can be found in a downtown movie theater in Boston. At this venue, these examples of circumvention were pretty wide-reaching, from a strong racial bias in the distribution of certain wages to superiors attempting to bully employees into accepting lower-paying positions.  In this instance, the company had developed a new work schedule for bartenders that involved them spending less time at the bar, and more time working at the cash registers. The bartenders are paid a lower hourly rate for time spent at the cash register, so they would make less money overall with this new schedule. Since no bartender would accept what was essentially a demotion, the general manager had private meetings with each of them to repeatedly ask them to accept a new contract. This led to an awkward situation in which the person I interviewed had to politely decline the requests to meet repeatedly since he had no one else to advocate for him.    In situations like these, it’s difficult for employees to advocate for themselves as their employers often see their positions as disposable. With jobs becoming increasingly more difficult to acquire, many workers would rather grin and bear such blatant violations of their rights than risk the possibility of being fired and losing a vital flow of income. Additionally, businesses often fail to tell workers their rights. In fact, some businesses may go out of their way to stop workers from knowing them, so a minimum-wage worker may not even know when they are supposed to advocate for themselves. Because of this, businesses can get away with far more than they should be able to.  To get a better sense of the shortcomings in minimum-wage jobs, I interviewed two workers in a large movie theater franchise. The main problems they mentioned were that their superiors (one in particular) had no checks placed upon them, and were able to exploit loopholes in the treatment of their employees. What we as a society need to do is provide legal safety nets in order to encourage them to report any shifty business in their hierarchies and explain to workers the steps they should take to advocate for themselves.  When asked what other workers in their position should do, the two theater employees I interviewed almost immediately said “Unionize. Get lawyers, and know your rights.” Here they referenced an organization that could provide legal representation for workers for a small fee. However, this is easier said than done, because if your employer believes that you and your coworkers are planning to form a union, they will sometimes try to fire you on the spot. That’s illegal, and if it happens you can sue, but lawsuits are costly and not a realistic option for a majority of minimum-wage workers. Some of the most important rights a worker has are the laws surrounding whistleblowers. If you, as an employee, alert the authorities to any kind of code violation in your workplace, you cannot be fired for doing so.  Just as important are the laws surrounding discrimination, which are well known but often not enforced. The main reason for this lack of enforcement lies in the fact that employees often go years without discussing things like income with their coworkers. It was in this way that the coworkers of the people I interviewed detected racial bias in wage distribution. After discussions, the employees found that the general manager was paying white managers more. To detect unequal wages due to race, interaction among coworkers should be encouraged. While some businesses may not condone this, it is important for all workers to be aware that such discussions are in no way illegal, and they cannot be legally punished by an employer.     As important as it is for employees to have legal support outside of the company they work for, it is just as important for workers to have a method of going above the heads of their superiors to resolve issues within their companies. Providing the contact information of upper-level management in a company is a good example of this kind of system. When the people I interviewed were asked if there was any way provided for them to go over the head of their general manager, they both said there wasn’t. One even said, “they probably don’t want us snitching on them.” Minimum-wage workers and interns in Massachusetts deserve to be advocated for. As a state and as a country, this is something that we as Americans need to work towards. The first step towards this would be the creation of unions for all workers, which would provide legal support for anyone facing off against massive corporations.  Also, laws should be implemented to ensure that employees have a way of reporting their superiors to upper management. And most importantly, we as people should respect the minimum-wage workers who we interact with daily during their jobs.  Read more… City News Teens should be warier of becoming gang-affiliated Gangs have a huge effect on our youth these days. Did you know that the average age to join a gang is just under 15? In Massachusetts, you can’t even get your permit at that age! Being gang-affiliated is life-changing, and it can ruin your life. According to Legal Match, being gang-affiliated is being associated with or close to a group of people, typically with three or more people who use certain names, signs or symbols to identify themselves; as a group or individually.  Teens are becoming awfully tribalistic when it comes to their fellow gang associates. People choose to associate with gangs because they think it gives them power, protection and easy money. It gives them a feeling of being untouchable and highly respected because they always have these gang members by their side. Often people assume that the more dangerous the gang, and the worse the reputation, the less likely you are to get bothered by anyone outside of the gang. Gang members often put on this act where the police and other gangs (unless they have a partnership) are their enemies. Ever heard the phrase “snitches get stitches?” That is street slang. It means that if you tattle on an ally to an opposing gang or law enforcement, the outcome will be a physical attack. This often makes people afraid of telling on other gang members. The information that’s kept from law enforcement due to this fear could be relevant to an open case, and withholding it could prevent a criminal from being brought to justice, or result in the withholder themselves going to prison. University of Washington researchers found that people between the ages 27 and 33 that were gang-affiliated were three times more likely to receive income from an illegal source and twice as likely to have been locked up. The same study also found that former gang members are twice as likely to have substance use problems and half as likely to graduate high school. Without a high school diploma, it’s extremely hard to get a job. Even if you did find a job, people without a diploma usually work for a very low income. High school dropouts make about $260,000 less than high-school graduates in a lifetime. That may not seem like a lot, but that’s $10,000 more than a quarter of a million. According to the Boston Police Department’s database, there are about 160 documented gangs, 5,300 total gang members, 2,800 active gang members and 2,500 inactive gang members in Boston. Many of these members may have joined gangs in high school. In 2010, there were approximately 772,500 young gang members in the United States. That was roughly 7% of the teen population then, so imagine the statistics now. Being affiliated with a gang doesn’t mean that someone is in a gang officially, it just means that they are associated with one. Authoritative figures shouldn’t punish or nag a teen for being associated with a gang, but teens should also know what or who they are getting involved with. There can be a real danger to even being gang-affiliated. For example, 15-year-old “Junior” Lesandro Guzman-Feliz was chased down into a bodega in Bronx, New York before being brutally killed by several men armed with machetes. This case immediately blew up and everyone came up with their own stories.  A popular theory was that it was an awful case of mistaken identity, but his untimely death was the result of gang affiliation. A show called “True Life: Crime” found that Guzman-Feliz had been seen hanging out with members from a New York gang called the Renegade Sunset, the Trinitarios rival gang. The Trinitarios are a well-known New York gang whose weapon of choice is a machete. Guzman-Feliz was on his way to hang with some friends when he encountered some members of the Trinitarios. Since he had been seen with Renegade Sunset he was considered a target, and members of the Trinitarios brutally murdered him. The point is, being gang-affiliated can be dangerous. Before you become comfortable hanging out with gang members, think of the outcome. Do not feed into peer pressure and be a kid while you can.   Read more…
Palm oil refinery plant After extracting palm oil, the crude palm oil is sent to palm oil refinery plant for refining. Through several refining steps to remove free fatty acids, natural pigments and odors, crude palm oil becomes refined-bleached-deodoried palm oil (RBD PO). The refined palm oil is close to colorless and transparent in liquid form. In addition, according to the needs of different users, palm oil can be further fractionated and processed to get palm oleic acid (PFAD), palm oil (referred to as OLEAN), palm stearin (referred to as STEARINE or ST). palm oil refinery plantAutomatic large scale palm oil refinery plant The business scope of Henan Doing Company: 1-20tpd small scale palm oil refinery plant 20-50tpd medium scale palm oil refinery plant 50-600tpd automatic large palm oil refinery plant 1-600tpd palm oil fractionation machine Palm oil brief introduction: Palm oil is extracted from palm pulp. In the fatty acid composition of palm oil, saturated acid accounts for 40% to 50%, and the rest is unsaturated acid neutralizing linoleic acid. Palm fruit is susceptible to decomposition of lipolytic enzymes before harvesting and palm oil production, and may also be affected during palm oil production process, palm oil storage and transportation. Therefore, the content of free fatty acids in hair palm oil is high, generally about 10%, and high up to 30% sometimes. But the phospholipid content in palm oil is very low. So Henan Doing Company recommend automatic palm oil refinery plant with Dry Physical Palm Oil Refining Method. Palm oil refining process: Palm oil refining process:Dry physical refining process of automatic palm oil refinery plant Dry physical refining process means that the degumming process don't use hot water, also called dry degumming process, which is suitable for oils with low phospholipid content (less than 200 mg/kg), such as palm oil, palm kernel oil, coconut oil and animal fats. The following is a brief introduction of Dry Physical Refining Process in automatic palm oil refinery plant. 1. Degumming Section: Heat crude palm oil to 30-50℃, add moderate phosphoric acid, and then the non-hydrated phospholipids transforms hydrated phospholipids. 2. Decolorization Section: Heat palm oil to 105-120℃, quantitatively add white clay and mix with palm oil. Utilization the adsorption of the white clay, the gums and the pigments are removed together. Besides, through leaf-filter's filtration process, thewaste white clay will be discharged. 3. Deodorization Section: Deodorization Section:Deodorization tower and internal structure of palm oil refinery plant The main machine in palm oil refinery deodorization section is combined packing deodorization tower that have two parts. The slender part above is deacidification part and the thick part below is deodorization part. Input high temperature steam to remove free fatty acids and odor components from the oil. In addition, the deodorizing portion can also lower the peroxide value of the oil. 4. Fractionation section: palm oil fractionationPalm oil fractionation machine The advantages of Dry Physical Palm Oil Refining: ◆ Low production consumption. Dry physical palm oil refining method needn't add alkli. ◆ Low cost of palm oil refining equipment. Dry physical palm oil refining method don't have neutralization process and washing process, so related machines are not required. ◆ No waste water and don't need wastewater treatment plant. Doing Company has been specialized in palm oil refinery plant for decades. And our palm oil processing machine's technology and quotation are both our advantages to compete with other companies. If you are interested in setting up a palm oil refinery plant, our professional engineers can help you design and install according to your capacity and requirements. Leave a message If you wanna to get more details about Palm oil refinery plant, you can send E-mail to . Or you can consult our professional engineers and specialized sales team by leaving a message in below form. We will contact you ASAP. You also can visit our factory in Henan, China. Leave a message
10 Types of Roof Trusses By Dawn Killough 12-03-2021 Tag Icon Roof trusses provide the structural support for a home's roof, supporting the roof deck material, insulation, and any loads that are placed on the roof. As a result, they have to be strong enough to withstand wind, rain, snow, and any other types of weather that your region might be prone to. Let's take a look at why it's important to know what kind of trusses your home has, how to determine what type of trusses your home may have, and what other types of roof trusses are available. Why You Need to Know Which Trusses You Have If you're going to invest in a remodel or addition to your home, you'll need to know what type of trusses are already inside. You'll want the roof's shape and slope to match your existing home and give the building a consistent look. If you plan to add attic space or a room in your attic, you'll also need to find what type of trusses were used. Roofing professionals may need to change them to attic trusses in order to add living space. Although most of them look somewhat similar, trusses come in a wide variety. The main differences lie in the shape of the truss and the location and configuration of the webbing. Understanding the structural interior of your roof makes it simpler to replace trusses and add new material to the home without sacrificing the home's aesthetic. How to Determine What Type of Trusses You Have Consider three key characteristics of your home to determine what type of trusses you may have: 1. Roof Shape: Roofs primarily come in three shapes: gable, hip, and gambrel. Once you know what shape your roof is, you can determine what types of trusses might have been used. For example, if you have a hip roof, you will have hip trusses. If you have gambrel-style roof, you will have gambrel trusses. 2. Ceiling Type: Your interior ceiling may also indicate what type of trusses were used. Having pitched or vaulted ceilings on the top floor of your home is a sign of scissor trusses. 3. Attic Space: If your attic space is finished or if you have a room in your attic, then your home should have attic trusses. If the attic is not finished, you can inspect the visible trusses to determine their style. Most Common Trusses for Homes Among residential roofing systems, some of the most common roof trusses include the following, which you can learn more about in the section below: • King and Queen: These post trusses are most commonly used on additions and new homes. The queen post variety can support longer spans than the king. • Fink Trusses: Commonly used in residential construction, fink trusses can carry larger loads compared with other varieties and are able to support new homes and additions. • Attic Trusses: Attic trusses make framing these attic spaces easier by including the wall framing in the truss. • Scissor Trusses: These trusses help homeowners achieve high, sloped ceilings, making them popular among modern homes with open floor plans. • Gable Trusses: This type of truss is usually placed at each end of a home with a gable roof to close off the attic space and provide support for exterior sheeting. 10 Types of Roof Trusses Both residential and commercial roofing systems have a wide range of possible roof trusses: 1. King Post A king post truss is the simplest variety of roof trusses with the fewest components—just two top chords, one bottom chord, one central vertical post called a king post, and two webbing chords. They help span short distances and are best for additions, garages, or other small spaces. 2. Queen Post A queen post truss is similar to a king post. Rather than having one centered vertical post, a straining beam connects two vertical posts that work well for new home constructions and larger home additions. 3. Fink Fink trusses typically have a role in constructing residential roofs. The webbing forms a W shape, allowing for greater load-bearing capacity that can come in handy during new home constructions and large additions. 4. Gambrel Gambrel trusses create a barn-shaped roof that is becoming popular with the resurgence of the farmhouse style. This uniquely tall design can add vertical space to a structure while spanning an entire house. 5. Attic Traditional trusses don't provide enough room for an attic living space, so attic trusses provide additional space under the roof. Attic trusses include the framing for the attic room in the truss design. They look similar to queen post trusses, but the posts are farther apart. 6. Scissor Scissor trusses have sloped bottom chords that create vaulted ceilings such as the ones in many open-plan homes. 7. Fan Used on medium-sized buildings, fan trusses are primarily made from steel and have a relatively simple style. The top chord is divided into smaller lengths, and the webbing is a combination of queen post and fink trusses. 8. Gable Gable trusses provide the end to a gable roof, a standard roof style for homes. It consists of two top chords, one bottom chord, and several posts set vertically for webbing. 9. Hip Hip trusses are used to create a hip, or hipped, roof. This style of roof slopes on all four sides, forming a pyramid shape that makes them a reliable choice in areas of high wind and snow. They have to be specially engineered with the correct slope and ridge location. 10. Flat Flat trusses are used for flat roofs, generally on commercial buildings. They are built similar to floor trusses and provide a lot of support. Searching for the right contractor for future roofing projects? Find a GAF-certified contractor* near you. Dawn Killough is a freelance writer in the construction, finance, and accounting fields. She is the author of an ebook about green building and writes for construction tech and green building websites. She lives in Salem, Oregon with her husband and four cats. Don't miss another GAF RoofViews post! If you drive down virtually any residential street in the U.S. or Canada, there's a chance you'll see at least one roof stained with black or dark brown streaks. Many assume the stains on their roof shingles are from mold, mildew, moss, or even tree sap. They're often actually caused by tiny bacteria called Gloeocapsa magma, AKA "blue-green algae". When you see the dark stains, you're not looking at the algae, but at the hard, dark coating it generates to protect itself from damaging UV rays. When snow and ice are in the forecast, it makes sense to start thinking about safety precautions for your family. Those preparations extend to your home as well—the formation of ice dams on your roof can cause significant damage. Why Attics Get a Bad Rep Attic's get a bad rep. Hot, unfinished, dark spaces, often hard to access, filled with insulation, ductwork, and old family heirlooms make attics an undesirable space. And scary movies like The Grudge certainly don't make attics any more appealing. What a shame. Perhaps if more people understood the potential that is just waiting to be unveiled, we'd have a very different viewpoint on these underserved spaces. Your attic needs to breathe. Poor ventilation may lead to excess heat and moisture and potential roof system degradation. The attic ventilation options that are best for your home will vary based on the style of your roof to ensure that each attic space has a balanced system for intake and exhaust. Balanced systems need to draw in fresh air and export hot, moist air. There's an old saying that it takes a village to support a family. In late October, it took 75 volunteers—including many GAF employees—working together with Habitat for Humanity of Tuscaloosa, Alabama to meet one goal: making a family's dream of homeownership come true. The best part? It took just five days. The evolving global pandemic and high number of natural disasters over the past year have continued introducing unique challenges to our neighbors, communities, and planet. In the midst of this, we have seen people all over the world stepping up to the plate to care for one another in new and creative ways. As we look back on 2021, we are touched by the impact we see others having and proud of what we've been able to impact ourselves.
Vitamin D Deficiency due to inadequate Sun exposure among women and children in UAE During my rotations in a Primary Health Care Centre in Abu Dhabi, it was noted that most of the women and children suffered from vitamin D deficiency. This prompted me to voice out on this issue in order to increase our awareness about this endemic ravaging our society. Vitamin D Deficiency The sun is a natural gift of God to man. The beautiful rays transcend beyond all boundaries, reaching every nook and cranny of the world.  For years on end, people have been benefiting from the refreshing rays of the sun. In the rural areas, people use the sun rays as a means of drying their clothes and preserving their foods whereas, in the urban areas, people benefit from its tanning ability, and electricity generation. The sun light is also essential for the photosynthesis of vitamin D in the body. If the major source of this vitamin is the sun which is natural and free without any cost, why do we then suffer the deficiency? History tells us that the Bedouin Arabs performed most of their activities outdoor, thereby absorbing most of the vitamin D needed by their body. In contrast, new generation of UAE women and children spend most of their time indoor, reducing their exposure to the sun light. This invariably increases their susceptibility to vitamin D deficiency since the sun is one of the major sources of this fat soluble vitamin. However, we can’t shy away from that fact that change in diet over the years has also contributed immensely to the deficiency of this vitamin. Vitamin D is made in the skin when 7-dehydrocholesterol reacts with UVB ultraviolet light at wavelengths between 270–300 nm which are present in sunlight. Depending on the intensity of UVB rays and the minutes of exposure, equilibrium can develop in the skin, and vitamin D degrades as fast as it is generated. The importance of Vitamin D can’t be overemphasized. It is essential to maintain skeletal calcium balance by promoting calcium absorption in the intestines, promoting bone resorption by increasing osteoclast number, maintaining calcium and phosphate levels for bone formation. It also allows proper functioning of parathyroid hormone to maintain serum calcium levels. In women, it is essential to prevent osteoporosis and osteomalacia. It is also suggested that vitamin D is very important when it comes to cancer prevention in women; the types of cancers we’re referring to are breast cancer, bladder cancer, ovarian cancer and colon cancer. In childen, Vitamin D plays an incredibly important role when it comes to growth and development. They usually complain of muscle cramps, difficulties in breathing and fragile bones (and skull) prone to fractures. Also if the child’s tooth are delaying and just won’t come out, the reason can be related to vitamin D deficiency. Ricket is one of the most common presentation in children with bowing of their legs. Dr Al Shehhi, Head of Rheumatology at Al Qassimi Hospital, Sharjah, said that out of 1,000 patients evaluated in the hospital, only 8 patients have normal levels of vitamin D. This is alarming! Also, Lab tests at the Shaikh Khalifa Medical City (SKMC) have revealed that 65% of females and 60% of males in the UAE suffer from Vitamin D deficiency. Therefore, to combat this menace, adequate health education is essential to increase the awareness of UAE residents on the importance of sun exposure and its adverse effects. Proper diets rich in vitamin D and dietary supplement could also go a long way in reducing the deficiency. They recommended an increase of daily oral intake of vitamin D to 800 IU for such women and 400 IU for children to ensure adequate vitamin D stores. Prevention is better than cure! Hidayat, Yetunde Ogunsola. 5th year M.B.B.S student. Gulf Medical University Ajman. U.A.E Previous Post Next Post Related Articles
Daily Devotional How Free Are We? April 7, 2019 - 12:00 am This Devotional's Hebrew Word Passover is also known as the holiday of freedom. We celebrate the fact that God set the children of Israel free. It’s a good time to also consider those who are not free. Indeed our freedom is something that we must never take for granted and that we should thank God for on Passover. Which leads to an interesting question: What is freedom, really? A story is told about Jews who approached a rabbi during the Holocaust, while millions of Jews were being held behind bars, and asked him, “Should we celebrate Passover this year? Are we free that we should thank God for freedom?” The answer the rabbi gave them was, “Yes, of course, you should celebrate Passover, since freedom is more than the ability to walk around and do whatever you want.” There are people who have been held in jail and have been free. Natan Sharansky, who as chairman of the Jewish Agency in Israel helped many Jews return to Israel, spent nine years in a Soviet prison. All that time, Sharansky exercised his freedom in any way he could. He did not lose faith in making it to Israel. And to symbolize his inner freedom, he disobeyed the Soviets at any chance he could. The day Sharansky was released from prison, he was blindfolded and told to walk straight ahead. As a truly free man, Sharansky walked in a zig-zag, just to celebrate his freedom. Millions of people walk around this planet with everything they need, including complete freedom – and yet, they live in a prison with invisible walls. They live for what others think of them. Or they say only what others want to hear them say. They have chosen a life path based on someone else’s dream, not their own. On Passover, we are required to see ourselves as though God personally redeemed us all over again. A good question to ponder during this time is: How free am I really? How much of my life is lived by choice and not by chance? How much am I influenced by the eyes of humans rather than the eyes of God? How able am I to be myself, say what I believe, and take actions that I know to be right? The best way to celebrate our freedom is by exercising it. One Response to How Free Are We? Leave a Reply
This Biotech Writes Long DNA Molecules From Scratch 12/04/2019 - 3 minutes Writing long stretches of DNA is difficult and expensive. The Austrian biotech Ribbon Biolabs is combining automation with enzymes to make the process easier, with the prospect of improving synthetic biology applications. ribbon biolabs synthetic biology synthesisMission: To synthesize DNA molecules with a length of up to 10,000 base pairs (the building blocks of DNA). This target is five times longer than what current DNA writing technologies can achieve. This technology could revolutionize the field of synthetic biology by making it easier and quicker to design organisms for manufacturing molecules such as proteins and biofuels. Reading DNA code is cheaper than ever before, as genome sequencing technology has made leaps and bounds in costs and efficiency. However, writing, or synthesizing, DNA has not seen quite the same progress. Today’s methods for automated DNA synthesis limit the sequences to 2,000 base pairs at most, which is below the average size of a gene,” Harold de Vladar, CEO of Ribbon Biolabs, told me. Trying to synthesize longer molecules than this takes a lot longer, and costs a lot more. Ribbon is developing a technology that writes genetic code using DNA-synthesizing enzymes, which could make the process faster than chemical synthesis, a method that’s been largely unchanged since the 1980s. Ribbon is also focused on automating the synthesis process to cut down the time and costs further. Founded in Austria in 2017, Ribbon raised over €1.5M last year to fund the development of its technology. The company now aims to produce its first 10,000 base pair DNA molecules by the end of this year. After this, Ribbon is ambitiously planning to produce DNA of up to 100,000 base pairs in length by 2020. Related Content ribbon biolabs synthetic biology What we think: Companies are racing to develop technology to synthesize long DNA molecules. By making this process as cheap and quick as genome sequencing, companies that achieve this feat could revolutionize synthetic biology. One example of the potential benefits of this technology include storing digital data in DNA. While traditional data storage technology lasts for around 20 years, DNA has the potential to last for thousands of years. Finding better ways of synthesizing DNA molecules could therefore make it easier to encode data in DNA. Another potential benefit is constructing genes and genomes. At present, researchers have to synthesize smaller DNA molecules and stick them together to make gene-sized DNA molecules. If Ribbon’s technology could create a whole gene in one go, this could make it easier to engineer organisms for specific purposes, such as manufacturing medicines and biofuels. Engineering organisms for better uses in synthetic biology is also the aim of many organizations. Other companies developing DNA synthesis technology include the French company DNA Script, and the UK biotech Evonetix. Related Content Improved DNA synthesis isn’t the only method of making engineering organisms easier. Another way is to simplify down the complex genome of organisms, such as bacteria, to more manageable sizes. Scientists based in Switzerland recently were the first to use computers to simplify the genome of a species of bacteria, which could make organisms better suited for manufacturing molecules. Images from Shutterstock Do you want to remove this advert? Become a member! Do you want to remove this advert? Become a member! You might also be interested in the following: Support Us Become a Member
What is a cloud based phone system? What is a cloud based phone system? A cloud phone system is a Voice over IP (VoIP) based business telephone platform that is hosted by a third-party provider such as OnSIP. Cloud phone systems typically offer advanced security and telephony features such as extension dialling, auto attendants, and conference bridges. Before the Cloud Phone System, in a Nutshell Business phone systems have been around for a long time — predating the internet, even. At a minimum, phone systems typically enable a business to have the main phone number with extensions to reach employees. Since phone systems became VoIP-based in the mid-'90s, companies have been able to run their own by installing telephony software on a server and connecting phones to the server over their IP network. But for a small or medium-sized business, running your own phone system requires IT overhead such as purchasing SIP trunking service or PRIs to connect the phone system to the Public Switched Telephone Network, maintenance, security measures, and custom configuration. (For more information about on-premise phone systems, read our blog on SMB VoIP System Options.) A "cloud based phone system" is a phone service delivered through the internet. Other synonymous terms include "internet based phone system" or more commonly "VoIP" (Voice over Internet Protocol). Both of these terms refer to a group of phones that work together, typically in a business setting, to route calls using an internet connection. The History of Cloud-Based Phone Services The telephone, the Internet, and Internet Protocol (IP) – without these three inventions, cloud-based phone systems wouldn't exist today. Invented in the 1870s, the telephone relayed calls with the help of an operator and a switchboard. The digital switching of calls began in the 1900s when AT&T created touch-tone buttons. This eliminated the need for a human operator. In the mid-1960s, the Advanced Research Projects Agency Network developed the very first version of the internet. It was used mainly for communications between the U.S. Department of Defense and the U.S. Military. It was only brought to public use in the 1980s when the Personal Computer (PC) became more popular. With these computers, people connected to the early internet via telephone lines. IP, which defines how information travels between two computers, was invented in 1972. Then, in 1989, the Hypertext Transfer Protocol and the Universal Resource Locator were invented. These inventions became the foundations of the internet as we know it today. Cloud-based phone systems or VoIP began in 1995 with a company called VocalTec, who pioneered what they called the InternetPhone. This was the first widely available Internet phone software. It allowed one Internet user to contact another and communication could be had through speakers and a microphone. It was developed to save money on long-distance and international phone charges. By 1996, Internet voicemail applications were created. In 1998, VocalTec went on to develop telephone-to-telephone and computer-to-telephone calling capacities for VoIP. Initially, the adoption of cloud-based phone systems was low, and this was mostly because users had to listen to advertisements before, during, and after their calls. The 1990s was a decade when telephone equipment manufacturers and telecommunication specialists started to utilise the new technology to their advantage. They began adding IP capabilities to their switches. Eventually, they developed software that allowed users to attach a VoIP adaptor to their phones. VoIP took off in the early 2000s. During this decade, three separate companies introduced VoIP switching software as add-ons for their routing equipment. Due to this accessibility, the number of VoIP calls skyrocketed to 25% of all voice calls. Call quality also improved with the availability of broadband Ethernet. In 2004, several companies that were full-service VoIP providers began to spring up. For a flat rate, they provided unlimited domestic calling services and free calling for those who were using the same VoIP provider. At this time, companies started taking full advantage of VoIP calling capabilities by creating unified communications bases. Today, the future of cloud-based phone systems is bright. It has become an integral part of many people's daily lives since they simplify many processes and offer a great degree of flexibility. There are sure to be further technological developments in the near future that will make cloud-based phone systems even more convenient, probably even indispensable. However, today's systems have the ability to deliver even more functionality for your company. If your communication strategy routes callers to specific departments, Ring Groups and ACD Queues provide loads of options. What Are the Benefits of a Cloud Phone System? Low Cost The primary benefit of choosing a cloud phone system over hosting your own is a reduction in time and money spent on running a phone system. It's a lot like choosing a hosted email service, such as Gmail, instead of creating your own. Not only can you skip purchasing the server(s) to host the phone system (and the SIP trunking or PRI service to connect it), you can skip all the time spent on configuration, upgrades, and security. Higher Security Believe it or not, VoIP hacking is prevalent if you're not careful with your phone system. If they can, VoIP hackers will take control of phone systems to make calls on your dime. Reputable cloud phone system providers are focused on running a secure service, so you don't have to. Ease of Use Cloud phone system providers offer administrative portals so that making changes (e.g. adding a phone for a new hire) can be done by any designated team member, with a few clicks. This is much more efficient for businesses than relying on a specific IT resource for complex configurations. Advanced Features & Integrations To stay competitive and relevant, cloud phone system providers focus on building calling features that help people communicate in a work environment. For example, a hosted phone system provider might offer a webphone that you can log into from anywhere, integration with your CRM for automated logging, or a click-to-call app so you can initiate calls with a click. Choosing the right cloud phone system for your business can help streamline processes and save the company resources. Scalability & Remote Working Finally, a cloud phone system will allow your phone system to scale with your business and support remote workers. Because the service is hosted on the internet, remote employees can register their phones from anywhere. Today, over 4 million people work remotely on a daily basis across the country. Businesses without a hosted phone solution often instead reimburse employee cell phone and home phone bills, which adds up. What are the pros and cons of cloud-based phone systems? Cloud-based phone systems are becoming more popular as the business world becomes more digitised and offers internet-based services. Cloud phones offer many benefits, ranging from cost to reliability, but before determining whether they are the best solution for your business, here are some important aspects to consider. Pros of cloud phones The main benefit of cloud-based phone systems is their low cost. Calls are conducted entirely via the internet, which means you are only being charged for your internet use, rather than for call minutes or extra phone service. Traditional phone systems can be expensive with multiple lines and add-on features like call transferring or voicemail. Most cloud systems offer the phone service with major features included in the service package, so you only have one monthly cost for all of your communication needs. Cloud phones are extremely reliable. These systems are hosted in multiple locations, which means that even if one server experiences issues, there is another server that can keep your communications running smoothly. This is called geographic redundancy, and it is a highly effective method for avoiding service issues and downtime. Flexibility is another major benefit of cloud-based phone systems. The world is moving rapidly toward an expectation of workers being available anytime, anywhere, and a cloud phone allows that functionality with the ability to route calls to your VoIP-enabled phone no matter where you are, as long as you have an internet connection. If your company is currently experiencing or expects to see major growth, a cloud-based phone system offers an excellent choice for easy and efficient scaling. A cloud management system allows you to add phone lines, extensions and advanced features easily without hiring a professional or paying for more equipment. This also goes for maintenance – all updates can be done in the cloud, saving you time and money. Cons of cloud phones Of course, there are some drawbacks to cloud-based phone systems that should be carefully considered. Cloud-based systems are entirely dependent on the internet and your connection. If you have spotty or unreliable service, a cloud-based phone system may not be the best option for you. A strong, consistent connection is required. If your business already owns many traditional phones, it may be cost-prohibitive to purchase all the necessary adapters or to purchase new IP phones. Consider the equipment you have and see if it makes sense to adopt that equipment into a cloud-based system. Cloud phones, being totally dependent on the internet, are susceptible to security issues like hacking and malware. Your connection must be protected. Furthermore, talk to the provider you're considering using and ask them to outline all the security measures they use to make sure your data is protected. VoIP and cloud-based systems also may not guarantee 911 calling support. Cloud-based phone systems are not tied to a single geographic location like traditional phones are, which makes it difficult to trace a 911 call accurately. An option is to enable Enhanced 911, or E911, in which customers set a physical address that shows up when they dial 911 on their cloud phone. However, not all cloud or VoIP providers offer this service, because to do so, they must cooperate with traditional phone companies (direct competitors) that have had limited access to the 911 network in the past. Ask the VoIP provider you're considering if it offers 911 service automatically, if it must be activated or if it is not supported at all. What are the factors that affect the quality of calls across cloud-based phone? In general, calls that run across cloud-based phone systems produce high-quality calling experiences. However, there are still some factors that can impact the quality of these calls. Often caused by electromagnetic interference, jitter is the interruption in digital signal traffic flow. It can cause audio clicking, monitor flickering, and data loss. Latency is the time delay of a packet of data reaching its destination. The amount of latency affects the usability of communication devices or interfaces. It is also the wait time for signals to travel across devices that are spread out geographically. Packet Loss Packet loss happens when transmitted packets fail to arrive at their destination. Data errors, jitter, and data corruption are its effects. Packet loss can be caused by system noise, inadequate signal strength at the destination, software corruption, hardware failure, or overburdened network nodes. This score is used in relation to voice testing processes, which range from 1 (worst) to 100 (best). It is based on the percentage of users who are happy with the quality of a test voice signal after it passed through a transmitter to a receiver. Mean Opinion Score The Mean Opinion Score measures the call's quality, with scores ranging from 1 (unacceptable) to 5 (excellent). More often than not, VoIP calls are in the 3.5 to 4.2 range. These areas of performance must be monitored so that the highest level of call quality is ensured and the number of dropped or interrupted calls is reduced. Why do I consider converting landline to VoIP? There are a lot of factors to consider when deciding the merits of a landline phone service versus a cloud-based one, especially for a business. Business phone services introduce a variety of needs and limitations that make weighing the differences between these two systems important. Unlike cloud-based phone systems, which use the internet to transmit your voice as data, traditional telephone lines rely on networks of copper wires, branch exchanges, and switches. This tried and tested system has been in place since the 1800s while the cloud-based phone system has only been in use since the 1990s. Traditional phone systems come with call forwarding, voicemail, call conferencing, and other features. However, those come at a cost. With cloud-based phone systems, those features are immediately available. Some features, such as voicemail to text, are unique to cloud-based phone systems. You can even integrate your cloud-based phone system with many third-party tools and software. You can't do that with a traditional phone system. Because a VoIP's service plan is based in the cloud, there is no need for bulky, expensive hardware. Installation is quick and just takes minutes. You only need a broadband Internet connection to start making calls, which makes the cloud-based phone system very portable. Through mobile applications, you can use your mobile phone as an extension of your office phone. The advantages of a cloud-based phone system: • Availability: The flexibility of the multi-device configuration in a cloud phone system will have multiple devices ring simultaneously, making you reachable even if you are out of the office. • Flexibility: Cloud PBX apps allow your employees to work from home or elsewhere and still take call transfers as if they were in the office. You can even have reception answer calls externally. • Scalability: You can start with a two-handset cloud system and scale accordingly as the business grows. This ensures you are only paying for what you need and can add handsets as needed. You can also reduce the capacity at any time. A cloud-based phone system operates in direct proportion to your monthly spending on telecommunications. • Multiple Sites: Cloud phone systems have the most benefits for businesses that operate in multiple sites. They let you break the shackles of geographic locations. Calls can be transferred from Sydney to Melbourne to Perth in a matter of seconds with internal extensions and no charge for the transfer of the call. How great is that? Multi-site offices save cost with free calls between all their sites. • Call Routing: Redirecting calls efficiently is an important part of the business. Our wide range of call routing features assures that a call is routed correctly to the right person. • Easy Relocation: When moving offices, you don't need to stress about leaving the exchange and losing your main number as you unplug your handsets from the old office and plug into your new office. How cool is that!? Quick Deployment: The turnaround time for the installation of the system is as quick as you need, eliminating the wait for appointment dates or technicians that sometimes don't even show up. Cloud based phone systems save money At first glance, a cloud based system may look like a poor investment. Most service providers charge a monthly fee. However, in-house costs are reduced to a simple hardware purchase. Costs involved with having to dedicate employees, resources, and providers all contribute to savings.  Traditional landlines are expensive to set up and host. What about internet-based systems hosted in-house? Even if you choose an internet-based business phone system (VoIP), your business costs will include organising to providers. With a cloud based phone system, the service provider will bundle costs into a rate that is determined per user. Budgeting becomes easier and scales with your business. Looking at an integrated business phone system solution can amplify the return you get from small business internet service providers. You can harness all the power of the cloud without paying for multiple hardware, software, and service components.  Cloud based phone systems maintenance Business phone systems are complex. Modern industry requires access to chat, email, voice calls, video conferencing, screen sharing, and e-meetings. However, integrating those various components across devices and locations can be costly in time and human resources. You want to get the most of your technology investments. Cut down on the number of software and hardware components and the labour that supports them.  With an in-house system, you have to have dedicated experts in each software and hardware element. This takes training and diverts your IT team's attention. Also, cloud based systems are more comfortable to monitor and maintain. Audits and replacements of entire systems are unnecessary.  With TPx, the initial costs of setting up your hardware are much lower than purchasing hardware systems, hosting and maintaining. You choose the UCx business phones that best suit your business phone needs. As a service provider, we take care of the management, monitoring, upgrades, and maintenance of your entire communication system.  Cloud based phone systems upgrading Upgrading should be done regularly to keep pace with technological advances. Cloud systems are the most comprehensive, cost-effective way of monitoring and upgrading your business phone solutions without downtime or expensive resources. Comprehensive communications with cloud based systems Unification Communications, as a Service, means that there is no redundancy or lags lost opportunities. Unlike other systems, cloud based systems optimiser force's ability to work collaboratively from multiple locations, across various devices. With TPx, you don't have to fuss with everyone's remote phone number or sending a confusing list of contacts to be mediated by a facilitator. It can be as easy as an app for end-users. Integration is more accessible to cloud based systems. As new applications and devices transform business workplaces, connecting them can be a cost-debilitating process in-house. By defraying the cost between clients, dedicated service providers can provide a hosted PBX that will allow maximum collaboration—no costs or downtime in upgrading your technology. Your system stays up-to-date and inclusive of the top communication technology at all times. Cloud based phone systems data security Whether it is sensitive communications themselves, the data shared through communication systems, or the access to endpoints security is of the utmost importance. The more we can connect, the more data moves. Having a cloud based system allows for better security.  Service providers like TPx offer dedicated data management and monitoring. In addition, by using the cloud, data backup is easier and safer. Important information is stored in various locations, so your data is protected against corruptions, loss, and disaster. And, disaster recovery is quicker. Cloud Phone System: Why you should switch! Economy: A cloud phone system is so much cheaper than the old phone system. There are no rental line fees, and this can save you a small fortune with you just realised your business has paid for this fee for all your landline numbers, during all years of operations. So easy to set up: with a cloud-based phone system you do not need expensive equipment. All you need is an internet connection and an IP Phone or softphones. You can get pre-configured IP phones from great brands such as Cisco (we provide you with many options of Cisco IP Phones), and once they have been installed, you are on your way! Then, all you need is an account with us. Flexible: with this phone system for the business, you have great flexibility. With the cloud-based method of connectivity, you can use your PC or mobile apps to make calls from anywhere, at any time. Better customer service: the cloud-based phone system offers your customer an excellent caller experience. Because it is so flexible, it reduces the number of missed calls from potential customers drastically, and this means that you do not miss out on sales. Phone Numbers: You can keep your existing number(s). In addition, you can also get new numbers. It is also possible to obtain international numbers so you can easily develop your business abroad. Cloud phone systems are suitable for all business in whatever sector. They can help a small business to operate in a manner like their larger competitors. Or if you are an enterprise organisation unify all your communication platform and manage your business telephony with a fully-featured system. A cloud phone system is a great fit for start-ups as well as we have a pricing strategy called: pay as you grow, depending on the number of users. And it will help your new business to look like an Enterprise. Frequently Asked Questions Virtual phone systems convert analog signals into digital and transmit them through the internet. They both provide phone functionality, but in different ways. The key differences between a PBX and virtual phone system are how calls are handled and how it connects to the Public-Switched Telephone Network. What are the different types of phone systems? There are three basic types of business telephone systems: KSU, PBX and VoIP. Scroll to Top
Skip to main content The importance of ventilation We have received Government guidance to remind customers about the importance of ventilating their homes during the pandemic. The virus is an airborne disease and so by opening your windows for a short while at regular times throughout the day you put yourself at lower risk to catching Coronarvirus. It's vitally important to open windows and keep doors open wherever possible when you have guests around in your home as this is the time when you are most as risk. Ventilation for the virus isn't the only benefit either. By opening your vents you can reduce condensation in your home which is a leading cause of damp and mould. Most newer windows will have a trickle vent similar to that pictured below, or you should have the opportunity to open your window by around 1cm and reclose the handle to offer a vented window. For older windows that don't have this option we'd encourage you to open your windows regularly to let air flow through and ensure they are open during high moisture inducing activities such as showers, washing clothes or cooking.
Wednesday, 5 August 2020 Surviving spring! The extreme drought conditions in most of the North Island and top of the South, wet conditions in Southland and the West Coast coupled with farmers holding more stock over autumn and winter (due to Covid-19) has created huge pressure on pasture supply. Supplementary feeds are in very short supply, and expensive, and winter forage crop growth has also been poorer than normal. Unless we have an extremely mild winter, there will still be considerable pressure on pasture supply moving into spring. Tactical use of nitrogen (N) fertiliser in late winter/ early spring, targeted so the extra grass grown satisfies a specific animal feed requirement, will be a cost-effective way of providing some much-needed feed. Given the vibrant microbiology in our grazed pasture soils, trial work shows little difference in speed and size of the N response to the type of solid N fertiliser, e.g. urea, DAP or ammonium sulphate, provided other nutrients are not limiting. This is because all forms of N are transformed by the microbes to nitrate within days to a couple of weeks. However, in cool and wet conditions there are other limiting factors that need to be understood when picking  the right N fertiliser. In late winter/early spring, after winter drainage events have leached much of the sulphate sulphur (S) present in soil, there may be a temporary lack of available S for the pasture as it responds to increasing daylength and temperature. Additionally, the crucial soil bacteria that mineralise S from soil organic matter for your pasture may not be fully active due to cold and wet conditions. Sometimes there will be a short-term S deficiency, which will limit the response of urea N, while ammonium sulphate applied at the same rate of N may give a bigger and longer response due to the added available sulphate sulphur. While we have no way of predicting where or when these limitations will occur for cold and wet late winter/early spring conditions, a product such as Ammo 31 or 36 is ideal as it combines both urea and ammonium sulphate and should be effective at minimising this risk. The reason for the better response to an Ammo product compared to urea under these conditions is likely due to two of the essential protein- forming amino acids, cysteine and methionine, which require sulphur as part of their structure. If the pasture plants can’t form enough of these amino acids, then the N response will be less. You may well ask that if an N fertiliser contains some nitrate, e.g. calcium ammonium nitrate or ammonium sulphate nitrate, will this work better in cold conditions? Trials have shown that even at soil temperatures between 2-6°C for three to four weeks after application, the extra response is barely measurable and not large enough to justify the much higher cost per kilogram of N in these products when applied to pastures.
Mysterious fast radio bursts ‘light up’ night sky in major discovery 11 Oct 2018 ASKAP antennas with the Milky Way overhead. Image: CSIRO Researchers in Australia have identified a treasure trove of mysterious fast radio bursts in deep space. Despite still having no idea what they are, Australian researchers using the Commonwealth Scientific and Industrial Research Organisation (CSIRO) telescope in Western Australia have discovered a record haul of fast radio bursts (FRBs). The mysterious bursts are believed to involve incredible amounts of energy – the equivalent of what the sun puts out in an 80-year period. In this discovery, the team has nearly doubled the known number of FRBs. Among the haul, the team reported in Nature, were the closest and brightest FRBs ever detected. The discovery of 20 in this latest survey was bolstered by the use of new technology found in the Australia Square Kilometre Array Pathfinder (ASKAP), which also confirmed that FRBs are coming from the other side of the universe, rather than within our own galactic neighbourhood. Over the course of billions of years, these bursts travel across the universe to us, but every so often they pass through a cloud of cosmic gas, slowing them ever so slightly. “Eventually, the burst reaches Earth with its spread of wavelengths arriving at the telescope at slightly different times, like swimmers at a finish line,” said the study’s co-author, Dr Jean-Pierre Macquart from the International Centre for Radio Astronomy Research. “Timing the arrival of the different wavelengths tells us how much material the burst has travelled through on its journey. And, because we’ve shown that fast radio bursts come from far away, we can use them to detect all the missing matter located in the space between galaxies, which is a really exciting discovery.” Localising the bursts Speaking of the technology involved, ASKAP engineer Dr Keith Bannister said this latest discovery was down to the telescope dish’s field of view – 100 times larger than the full moon. When multiple dish antennas point at different parts of the sky, this bumps up the field of view even more, to an equivalent of 1,000 times the area of a full moon. The team will now look to pinpoint the locations in the sky where the FRBs came from. Lead author of the research, Dr Ryan Shannon, explained what this will mean. “We’ll be able to localise the bursts to better than a thousandth of a degree. That’s about the width of a human hair seen 10 metres away, and good enough to tie each burst to a particular galaxy.” Colm Gorey was a senior journalist with Silicon Republic
This summary is a very fast-paced account of Bible History as related in the historical books of the Old and New Testament. The author purposely omitted many details in the preparation of this summary in order to keep the thread of historical development in full view. It is the author's earnest hope that this work would be helpful to someone who wishes to have a reasonably complete overview of Bible History before reading the actual books. It is best to read this summary with the timeline given under the charts menu. The button below is a link to the Timeline. A.  From Adam (Before 9000 B.C.) to Abraham (2100 B.C.) The book of Genesis records that God created the heavens and the earth in six days. He created the whole world - the seas, the land, the sky - and all its inhabitants: the fishes of the sea, the animals on land as well as the birds of the air. He also created the first man and woman - Adam and Eve - to His own image and likeness, and placed them in a paradise of pleasure, where they lived in a state of happiness. (Gen 1-2) God gave Adam and Eve dominion over all His creatures on earth, and told them that they could eat the fruit of any tree in paradise, except that of the one that stood in the midst of the garden – “the tree of knowledge of good and evil,” for on the day that they eat of that fruit, they shall die. But a cunning serpent tempted them and said, “No, you shall not die if you eat the fruit; instead, you shall be as Gods, knowing good and evil.” Adam and Eve listened to the serpent, ate of the fruit, and broke God’s command. Because of their disobedience God banished them from paradise, and condemned them to a life of hardship and eventually to die. But God had mercy on them, and promised to send them a Savior who would one day reconcile them again with Him. (Gen 3) Cain and Abel were among the first sons of Adam and Eve. Although both offered sacrifices to God, God favored Abel, but not his wicked brother, Cain. Being exceedingly jealous of Abel, Cain killed him. For this sin God cursed him and made him a vagabond and a fugitive upon earth. (Gen 4) Like Cain, many of the other descendants of Adam and Eve also turned away from God; so much so that in time the wickedness of men filled the earth with sin and vice. Therefore, God resolved to punish the world by a universal Deluge (probably around 2500 B.C.). However, He found one man, Noe, to be pious and just, and wished to spare him and his family from the calamity that was to come. So, God instructed Noe to build an ark, in order to save him and his family, as well as all the animals in it, from the great flood. The rain fell for forty days and forty nights, until the waters rose 15 cubits (approx. 27.5 ft) above the highest mountains. When the waters finally subsided Noe built an altar to thank the Lord who, in turn, blessed him and his sons and promised not to destroy the earth again with flood. (Gen 5-9) Noah's Ark A painting by Edward Hicks, 1846 Image source link: commons.wikimedia.org After the flood Noe’s family increased and multiplied, and became as wicked as they were before the Deluge. At the time they spoke just one language. As they had a huge labor force, they thought of building a tower so tall that its top would reach heaven. When God saw what they were doing He confounded their language. Not being able to understand one another, they were forced to abandon the Tower that they were building. The unfinished tower has since been called the Tower of Babel, or Confusion.  Because they could not understand one another, Noe’s families spread out in different directions, but they did not abandon their wicked ways.  Many lost their faith in the true God, and had degenerated into idol-worshiping nations. But one kept his faith. His name was Abram, which God later changed to Abraham. (Gen 10-11) B.  From Abraham to the Conquest of Chanaan (2100 B.C. – 1350 B.C.) God did not want the true faith and the hope in the promised Redeemer to be lost entirely from the earth, so He chose Abram to be the father of what would be His Chosen People. God changed Abram's name to Abraham and promised him that he would be the father of a great nation. This is how Abraham’s descendants – the Hebrews – was chosen to be the nation, out of all other nations, in whom the true religion was to be preserved. (Gen 12) To test his loyalty God ordered Abraham to take his only son, Isaac, to Mount Moria, and there to offer him up as a holocaust. Abraham obeyed, and made Isaac carry the wood as they went up the mountain together. Upon reaching the summit, Isaac willingly laid down on the wood to be offered as a burnt sacrifice, but God sent an angel to stop Abraham from killing him. Theologians see in Isaac a figure of our Savior who, in obedience to His Father, carried the wooden cross to Mount Calvary to offer Himself for our redemption. Pleased by Abraham’s willingness to sacrifice even his only son as a burnt offering, God made another special covenant with him saying, “I will multiply thy seed as the stars of heaven … and in thy seed shall all the nations of the earth be blessed” (because the Messias will come from his seed or lineage.)  (Gen 22) Isaac became the father of Jacob, whose name was later changed by God to Israel (Gen 32:28-29; 35:10). Jacob (or Israel) in turn had twelve sons who became the fathers of the twelve tribes of Israel.  In Biblical terminology Abraham, Isaac, Jacob and his twelve sons are called the Patriarchs (or Fathers), for they were the men who established the nation of Israel. These are the twelve sons of Jacob: Reuben, Simeon, Levi, Judah, Dan, Naphtali, Gad, Asher, Issachar, Zebulun, Joseph, and Benjamin. Out of the descendants of Levi (the Levites) will come the priests of Israel. Joseph, who was sold by his brothers to slave traders and brought to Egypt, later became Chief Steward and Vizier of all Egypt. At that time Egypt was a very powerful kingdom (See Map of the Egyptian Empire). The Pharaoh granted him this position of authority because, by his ability to interpret dreams, he predicted the coming of seven years of famine, and saved the country from dreadful hardship. Indeed, the Pharaoh called Joseph “Savior of the World” (Gen 41:45), which made him – just like Isaac – a figure of Christ. Joseph, Overseer of the Pharao's Granaries A painting by Lawrence Alma-Tadema (1836-1912) Image source link: commons.wikimedia.org The famine, which also broke out in Chanaan, drove Jacob and his family to Egypt. The kind-hearted Joseph forgave his brothers and invited them to settle there in Egypt with their father (Gen 37-41).  It is probable that the Israelites entered Egypt in 1876 B.C. The Jews prospered and were happy in Egypt. They learned new trades and skills from the Egyptians. But, after Joseph’s death, the Egyptians feared that the Israelites were becoming too numerous and too powerful. So, they persecuted and burdened them with excessive labor. Eventually God talked to Moses, son of a Hebrew slave and adopted child of the Pharaoh’s daughter, at Mount Horeb. There He commissioned him and his brother, Aaron, to lead His Chosen People back to Chanaan. The Pharaoh at first would not let the slaves free, but God sent dreadful plagues to make him change his mind. To overcome the Pharaoh’s obstinacy, God finally sent an angel to slay all the first-born children in the land. The angel killed the first-born of the Egyptians, but not the Israelites. This is because the doors of Israelite homes had been sprinkled with the blood of the paschal lamb which they ate at supper. That the first-born of the Israelites were saved by the blood of the paschal lamb prefigures salvation from death by the Blood of Christ, whom we eat in the Holy Eucharist. The Pharaoh finally allowed the Hebrew slaves to leave, but regretted it. This happened around 1446 B.C. The Pharaoh gathered his army and pursued the Israelites to the banks of the Red Sea. The Israelites were terrified by the specter of being slaughtered by an angry King. Moses stretched forth his staff, and the sea was divided, with the waters standing like a wall on their left and on their right, allowing the Israelites to cross the sea on foot. When they reached the other side of the sea, the Pharaoh and his troops went hurriedly after them. But Moses stretched forth his staff once more and the waters of the sea returned to their original place, drowning the Pharaoh, his horses and all his men (Ex 1-14). Crossing of the Red Sea A painting by Juan de la Corte (1580-1663) Image source link: commons.wikimedia.org Under the leadership of Moses and Aaron the Israelites traveled through the desert and, after 45 days since they left Egypt, they arrived at Mount Sinai. How do we know that it only took the Israelites 45 days to reach Mount Sinai? Well, Ex 19:1 said that the Israelites came to the wilderness of Sinai on the third month since they departed Egypt. Now, this does not mean three months after they departed Egypt. The third month starts after two months, so their travel time to Mount Sinai is at most two months or 60 days. However, Ex 12:6 also stated that they actually departed on the 14th day of the first month. Therefore, their total travel time from Egypt to Mount Sinai is only about 45 days. While in their journey the Israelites grumbled and murmured because they were hungry. They didn’t have enough provisions to eat.  But God sent quails for their meat and made manna appear on the ground every morning for their bread to eat. This helped to convince them that God knew how to take care of His chosen people. Upon reaching Mount Sinai God renewed His covenant with them and gave them the Ten Commandments on two tables of stone. In addition God gave them civil, moral and religious laws and precepts, which showed them how they should conduct their lives. But because Moses was often long absent talking to God in the mountain, the Israelites soon forgot God’s blessings and His Commandments. Worse, they fell into idolatry and, led by their priest Aaron, they made for themselves a golden calf to worship as their god (Ex 16-32). At Moses’ intercession the Israelites were reconciled with God. They made offerings for the construction of the Tabernacle (a meeting tent) and its furnitures. They also built the Ark of the Covenant, which housed the two tablets of stones on which the Ten Commandments were written (Ex 33-40; Lev 1-10). But in punishment for their serious sins God delayed their entry into Canaan - the Promised Land. The Israelites remained one year waiting in the wilderness of Sinai before they were permitted to move again toward Canaan. Then they wandered thirty nine more years in the desert before they reached the Promised Land. These thirty eight years were full of murmurings and revolt by the Israelites against God. Yet, God still protected them and sustained them during their long stay in the wilderness. He dropped manna from heaven for bread, sent quails for meat, and let water flow from the rocks for their drink. He also protected them against serpents and foreign invaders. (Num 10-21) Because of their long sojourn in the desert, many of the older Israelites – including Moses – did not make it to Canaan (or Palestine). Moses died before they entered Canaan. The death of Moses is recorded in Deut 34:5.  The text shows that Moses saw the Promised Land, but he himself did not get there. Although the authorship of the first five books of the Bible (aka the Pentateuch) is attributed to Moses, it is clear that the last chapter of Deuteronomy, which records his death, was only added by others (possibly by Joshua). It is also possible that other parts of the Pentateuch had been edited by others. Still, it is not incorrect to attribute the entire Pentateuch to Moses as its original author. After Moses died, God chose Josue, of the tribe of Joseph, to lead the huge army of the Chosen People in the task of occupying Palestine. Under his leadership the Israelites crossed the river Jordan and miraculously laid siege to Jericho. The conquest of Jericho was miraculous because the city was heavily fortified by huge walls. The Bible records that the walls collapsed at the sound of the trumpets and the shouting of the Israelites (Jos 6). Of course, the sound of the trumpets and the noise of the Israelites did not physically cause the walls to collapse, but by God’s will and instruction these were the signals for the miracle to begin, and for God’s Power to show its might. After the fall of Jericho the Israelites conquered western Canaan, too, but after many battles. When the conquest of Canaan was complete, the land was divided among the various tribes of Israel (Jos 2-22). See the Map of Canaan below to see how the Promised Land got allocated among the tribes. Tribal Allotments of Israel.png Tribal Allotments of Israel Picture # 39 of Bible Atlas by Access Foundation Image source link: Bible Atlas Note that the tribe of Levi does not appear in the map. Although the Levites received cities in different parts of the country, they were not permitted by the Law of Moses to become landowners because, as priests, “the Lord Himself is their inheritance” (Deut. 18:2). Joseph’s name is not in the map either. But Ephraim and Manasseh were both sons of Joseph (Gen. 48:1), so together they constituted the house of Joseph. With the conquest of Canaan ends the early history of the Jews. In a sense their history is also a figure of our personal salvation history. For, their deliverance from the bondage of Egypt represents our liberation from the slavery of Satan; their arduous journey in the wilderness corresponds to our difficult pilgrimage in this world; the manna they received  in the desert is a figure of the Holy Eucharist, which sustains our souls; and the Promised Land where they settled after many battles signifies Heaven, our final home, which we attain by a series of battles against the world, the flesh and the devil. C.  The Rule of the Judges (1350 B.C. – 1051 B.C.) Although Canaan had been divided among the various tribes of Israel, some pagan tribes remained in their respective territories.  When Joshua died, the Israelites were left without a leader, and they became vulnerable to their pagan neighbors in and around Canaan. The association of the Jews with the heathens also led to intermarriages with the Gentiles that in turn resulted into their forgetfulness of their covenant and their subsequent fall into vice and idolatry. As a punishment for their repeated sins of infidelity and immorality, God allowed the Israelites to be subdued from time to time by their enemies.  But whenever they repent, God would send them a strong military leader who would rescue them from their enemies and rule over them. Because these military rulers also presided over legal hearings, they were more popularly known as “Judges.”  For three hundred years the people of Israel lived under the rule of these Judges. Among the most famous of the judges were Gideon, Samson, and the pious Samuel. (Jgs 6-8, 13-16; 1 Kgs 1-7)  The charming story of Noemi and Ruth (great grandmother of the future King David) also happened during the time of the Judges. (Ruth 1-4) D.  The Reign of Kings to the End of the Babylonian Captivity (1050 B.C. – 538 B.C.) Because of increasing attacks by the Philistines from the west and the Ammonites from the east, the Jews wished to unite the various tribes into one kingdom. So, they asked the aging Samuel (the last of the Judges) to appoint a king for them. Under God’s guidance, Samuel chose Saul, who became the first King of Israel. He reigned for 40 years (Acts 13:21) from 1050 B.C. to 1010 B.C. Because of his disobedience, he was later rejected by God and succeeded by David, a mighty warrior. Even as a young lad David had slain the giant Goliath. As a king, he expanded his kingdom by many victorious battles against his enemies (1 Sam 8-31). Like Saul, David reigned for 40 years (2 Sam 5:4), from 1010 B.C. to 970 B.C. David was more than a king. He was also a song writer. Under God’s inspiration, he composed many sacred songs, or psalms, that contained prophecies about the coming Messiah (2 Sam 1-7).  For example, he hinted that the Messiah would be King and Priest (Ps. 110:4), that he would be the Son of God (Ps 2:7), that he would be betrayed by one whom he trusted (Ps 41:9), that his hands and feet would be pierced (Ps 22:16), and that he would rise from the grave (Ps 16:10). To properly understand the verses in the Psalms, it is important to keep in mind that David was a prototype of Christ. Therefore, although the psalms were often written in the first person, the verses should be read as if Christ was the one speaking, and not David. David’s son, Solomon, succeeded him to the throne, and reigned also for 40 years (1 Kgs 11:42) from 970 B.C. – 930 B.C. Solomon was a wise and peaceful ruler. His reign was characterized by peace and great prosperity, and his wealth was unsurpassed by that of any other king.  He built the temple in Jerusalem, which David originally wanted to build.  The temple was lavishly decorated with gold from Ophir (a place rich in gold), and was the only temple in the land. The Sanctuary or Holy of Holies, where the Ark of the Covenant was placed, was the most sacred place in the temple, and only the high-priest was allowed to enter it once a year. The Jews were not allowed to offer sacrifices anywhere else except in the temple. This magnificent edifice was the pride of the Jews. But Solomon did not persevere in virtue. He married many pagan wives and later fell into the practice of idolatry himself (1 Kgs 3-11). King Solomon Sacrificing to the Idols A painting by Sebastien Bourdon (1616-1671) Image source link: commons.wikimedia.org After Solomon’s death, his son Rehoboam became king, but the northern tribes rebelled against him and established their own kingdom in the north. Only the tribes of Benjamin and Judah in the south remained faithful to him. Thus, the kingdom of the sons of Jacob had been divided into two. The northern kingdom, called the Kingdom of Israel, was ruled by King Jeroboam, and had its capital in Samaria. The southern kingdom, known as the Kingdom of Judah, continued to be held by King Rehoboam, and had its capital in Jerusalem (3 Kgs 12). The northern tribes abandoned the religion of their fathers and built their own temple in Samaria, where they practiced various forms of idolatry. During this time God sent them holy and enlightened men, called Prophets – such as Elijah, Elisha, Amos and Hosea – to call them to repentance. The Book of Jonah narrates that God also sent the prophet Jonah to preach repentance to the pagans in Nineveh. The pagans repented and Nineveh was spared from God’s wrath. But the Jews in Israel continued their wicked ways. As punishment for their transgressions, God allowed them to be conquered by a foreign power. King Shalmaneser V of Assyria started the attack on the Kingdom of Israel and laid siege to its capital, Samaria. Upon his death his brother, Sargon II, continued the siege, and in 722 B.C. destroyed the northern Kingdom of Israel, and carried the Jews off as captives into Assyria. An episode of this captivity was written in the Book of Tobit, but nothing else was heard of the Jews afterwards. They had since disappeared from history as the “lost” tribes of Israel.  Some Assyrians moved to Israel and intermarried with the few remaining Jews in northern Palestine, which became the Samaritans – a people of mixed blood, culture and religion (1 Kgs 17-22; 2 Kgs 1-18). The foregoing discussion is very fast paced and omitted many details.  The total length of time that the Kingdom of Israel existed, from the division of Solomon’s kingdom to the fall of the northern Kingdom of Israel, is about 930 – 721 = 209 years. Therefore, there were actually several kings who reigned in the Kingdom of Israel after the reign of King Jeroboam, such as Nadab, Baasha, Elah, etc. One king in particular, Ahab, will be remembered for his wickedness (1 Kgs 21-22), and his queen Jezebel for her harlotries and sorcery (2 Kgs 9:22). Meanwhile in the south the tribes in the Kingdom of Judah maintained an independent but unstable existence. God also sent them holy prophets, – such as Isaiah, Jeremiah, Baruch, Joel and Micah, Zephaniah, Nahum, Ezekiel and Habakkuk, – to warn them of the consequences of their sins and their need for repentance. In the end God also chastised the Kingdom of Judah for its repeated sins. For, after the Babylonians destroyed Nineveh and defeated the Assyrians in 612 B.C., the powerful Babylonian army also went to Palestine in 607 B.C. and, under the command of Nebuchadnezzar II, King of Babylon, they conquered the Kingdom of Judah and pillaged the city of Jerusalem. Then in 587 B.C. Nebuchadnezzar destroyed Jerusalem and its temple altogether, and carried off most of the Jews into captivity in Babylon (2 Kgs 25:10-21). As in the Kingdom of Israel, there had been a long line of kings in the Kingdom of Judah until its fall in 587 B.C. There were good and bad kings. King Uzziah began as a good king, but later tried to usurp priestly functions and was later punished with leprosy (2 Kgs 15; 2 Chr 26). On the other hand, King Hezekiah was so pious that his kingdom was blessed with prosperity and even won a battle against the powerful Assyrian army led by Sennacherib (2 Chr 29-32). Judith, a beautiful and wealthy widow during the wicked reign of King Manasseh, was neither a queen nor a ruler of Judah, but the book of Judith immortalizes her fidelity and heroism, which saved the nation from an earlier destruction by Holofernes, the mighty general of Assyria. The history of Judah had been a series of sinfulness and repentance, with God saving them whenever they repent. After their fall in 587 B.C. Judah became part of the Babylonian Empire (See Map of the Babylonian Empire). The Jews then became captives in Babylon for five decades until Babylon itself was conquered by the Persians in 538 B.C. (2 Kgs 14-25). This is known in Bible History as the “Babylonian Captivity.” The Lamentations of Jeremiah spoke of the deep sorrow he and other Jews felt due to the destruction of their once great city: Jerusalem. At a more profound level the Lamentations allude as well to the sorrows of our Suffering Savior, and to man’s sad condition when separated from God. The inspiring stories of the three young men in the fiery furnace and of Daniel in the lion’s den, both took place during the time of the Babylonian Captivity (Dan 2-14) E.  From the End of the Babylonian Captivity to the Coming of Christ (538 B.C. – 1 B.C.) In 538 B.C. Babylon became a part of the Persian Empire. (See Map of the Persian Empire.) Cyrus, King of Persia, gave permission to the Jews to return to their own country, but under the rule of a governor appointed by him. Zerubbabel – who happened to be one of Christ’s ancestors (Matt 1:12; Luke 3:27) – was the man appointed by King Cyrus to govern and lead the first group of Jews that would return to Judea. In 537 B.C. more than 40,000 Jews (Ezr 2:64) returned to their homeland. This was actually only a fraction of the Jews in Babylon who returned to Judea. Many of the Jews became so comfortable with their life in Babylon that they decided to stay there. The mission of the returning exiles was to rebuild their temple and their city. King Cyrus was kind to them. He did not merely allow them to go back to Palestine, but even returned the vessels of the temple that were stolen by Nebuchadnezzar, and provided funds (gold and silver) for the reconstruction of the temple. But it was not until the reign of King Darius I of Persia that the temple of Jerusalem was completed. The temple was dedicated in 516 B.C. Although the new temple was less impressive than the original temple built by King Solomon, the Prophet Haggai prophesied that it would be a more glorious house than the former, because the “Desired of All Nations” (Christ) would one day enter it (Hg 2: 8-10). Another prophet during this time was Zechariah, who prophesied that Christ would one day enter Jerusalem riding on a donkey (Zach 9:9). Jewish Exiles Returning to Jerusalem Illustration from a Bible Card, 1898 Image source link: commons.wikimedia.org King Darius I was succeeded by Xerxes I (also known in the Bible as King Ahasuerus). At the time there were still many Jews remaining in Babylon. It was exactly during the reign of King Ahasuerus that the beautiful story recorded in the book of Esther happened. According to the narrative, King Ahasuerus of Persia married the beautiful young maid, Esther, not knowing that she was a Jew. By advice of a jealous prince, Haman, Ahasuerus was about to massacre the Jews in Babylon when Esther came to the rescue. With prayers and fasting Esther boldly revealed her identity to her husband and disclosed the conspiracy unfairly plotted by Haman against the Jews. Under the influence of his good and sincere wife, King Ahasuerus abandoned the plan to kill the Jews and punished his adviser, Haman, instead. Theologians view Esther’s successful petition before the king as a foreshadowing of the Blessed Virgin Mary’s powerful intercession before her divine Son and God the Father. Queen Esther A painting by Andrea Castagno, 1450 Image source link: WikiArt.org Eventually King Artaxerxes I succeeded Ahasuerus to the throne. Since there was now a temple again in Jerusalem, Ezra, a holy priest and scribe of the Law persuaded and led a new group of exiles to return to Judea. He was given power to appoint judges and magistrates in Jerusalem, but his real desire was to see a Jerusalem free of paganism. He worked to prevent the Jews, many of whom had married pagan wives, from relapsing back to idolatry. He gathered all the books that had been written to form what is now known as the Hebrew Bible, and used this to re-establish readings and Divine services in the Temple. Prophets during this time period were Malachi and Joel. When his cup-bearer, Nehemiah, reported to him the sorry state of Jerusalem’s walls and how vulnerable it was to attack by the surrounding Samaritans, Artaxerxes I commissioned him to rebuild its walls and to fortify the city’s defenses against its enemies (Neh 2, 3 & 4). Years afterward, in 332 B.C., the Persians lost their power to Alexander the Great of Macedonia (Greece), so the Jews fell under Greek rule. Fortunately, Alexander the Great was also kind to the Jews, so peace continued in Judea. (See the Map of Alexander the Great's Empire.) After the death of Alexander the Great in 323 B.C. there followed a power struggle among his four generals, and the Greek Empire was divided between them. The Kingdom of Judah fell under the control of one of the generals of the empire, who happened to be Ptolemy, the King of Egypt. The Egyptians therefore ruled over Judah for about one hundred twenty-five years, from 323-198 B.C. It was one of the Egyptian kings who later requested the high-priest in Jerusalem to work on a Greek translation of the Holy Scriptures. The result was the Greek Bible, known as the Septuagint (LXX). See History of the Septuagint. In 198 B.C. the Egyptians lost control of Judah to another one of Alexander’s generals – the Seleucids of Syria. During this time the people of Judah had already repented of the sins of their ancestors that had brought them so much suffering, and they promised never again to fall into idolatry. So, when Antiochus IV, King of Syria, forced them to worship idols (168 B.C.), they valiantly resisted the king’s order. Following the example of the seven brothers and their mother (2 Macc 7:1-42), they chose martyrdom rather than disobey God. Martyrdom of the Machabees A painting by Giustino Menescardi (1720-1776) Image source link: commons.wikimedia.org Under the leadership of their high-priest Mattathias and his sons, the Jews steadfastly continued their rebellion against the Syrians until they freed Palestine from Syrian rule. They even extended their campaign to recover parts of the former Kingdom of Israel, but their victory was short-lived. In 67 B.C. two brothers struggled for control of the Kingdom of Judah – Hyrcanus II (the high-priest) and Aristobolus II (the head of the army). Overpowered by his brother, Hyrcanus sought the aid of the Arabs. But the Romans took notice of the situation and intervened. They asked the Arabs to withdraw and, in 63 B.C. under the command of Pompey, they invaded Palestine and annexed it to the Roman province of Syria. (See Syria in the Map of the Roman Empire). The Romans, now in control of Palestine, gave the Jews much latitude in the practice of their religion and the management of their internal affairs. The Sanhedrin, or the Council, became the supreme religious and judicial court for the Jews. It was presided over by the high priest, and consisted mostly of the “ancients” and the “rabbis” (or lawyers). It had great authority that extended to all Jewish communities throughout the empire. It even had its own soldiers or police, but it could impose the death penalty only by consent of the Roman procurator.  But the two groups that had the greatest influence in the life of the Jews under the Roman Rule were the Pharisees and the Sadducees. The Pharisees did not all come from the priestly class, but they were the conservatives who insisted on the strict observance of the Law. Although they helped protect the true religion from pagan influences, they tended to become separatists, aloof, and too legalistic.  On the other hand, the Sadducees were more liberal in their conduct and in their views. While they respected Moses and the Law, their own materialism and rationalism shaped their beliefs to the point of heresy. For example, they denied the immortality of the soul, the resurrection of the body (Acts 23:8), and the fact of God’s providence. The Jews did not like paying taxes to a foreign power and being subjected to foreign rule. So, riots and rebellion often broke out in Palestine, which were ably suppressed by the powerful Roman army.  In a strategic move to control Palestine more effectively the Roman Senate in 40 B.C. appointed a clever but cruel man, Herod, to serve as king of Palestine. This was Herod the Great, the same king who is infamously remembered in history for slaughtering the Holy Innocents during the infancy of Christ. He bribed the Romans to make himself king of Palestine. Although he was a Jew by religion, he was actually not a Hebrew but an Idumaean (a Roman term for an Edomite). Thus, the “scepter” of Judea was no longer in the hands of the tribe of Judah when Christ came, fulfilling a prophecy made by Jacob before his death (Gen 49:10). Actually, King Herod was only a “puppet,” because the supreme political power still remained with the Roman governor.  But, to empower himself more fully, Herod removed all executive authority from the Sanhedrin and transferred them to himself, thus reducing the Sanhedrin into a purely religious body. This was the political and religious climate of Israel when Christ was born, which was around 1 B.C. Although the Jews had returned to the true religion, moral corruption, greed, and political ambition continued to plague them as they did most of humanity. The Jews had long waited for the Messiah, but due to their long history of being continually subject to foreign rule, many of them thought of the Messiah as someone who would liberate them from foreign domination, rather than someone who would redeem them from sin.
In this course, you’ll learn common ways of keeping good computer hygiene to reduce the threat of ransomware. • List best practices that reduce the likelihood of being affected by ransomware All computer users 1. How Keeping Your Operating System Up-to-date Thwarts Ransomware 2. How Keeping Unclean Practices off of Systems Used for Work 3. How Using Good Password Behaviors Can Prevent Ransomware 4. How Being Careful with USB and External Media Can Prevent Ransomware 5. How Being Careful with Public Phone Chargers Can Prevent Ransomware 6. How Using Only Trusted Software Can Prevent Ransomware
Facebook Twitter Blood Sugar Spikes After Exercise Diabetes, Exercise and Blood Sugar Control What makes blood sugar spike AFTER exercise? Intense workouts can spike blood sugar levels. When someone with diabetes exercises, often the concern is about making sure they have enough energy to complete their program. That energy is in the form of glucose or "blood sugar" that comes from food and drink. Without adequate amounts of glucose, they may experience something called "low blood sugar" or when it's more serious, "hypoglycemia." Effects can range from mild dysphoria to more serious issues such as seizures or even unconsciousness. Here's how things normally work. When we exercise, our muscles use blood sugar for energy throughout the workout. The longer and more intense the program, the more blood sugar you need to make it through. That's why we always suggest clients eat 1-2 hours before a training session, so they have the energy or blood sugar to finish. But that's not what happened to a recent client. After his workout, he took a reading of his blood sugar and discovered it was much HIGHER than when he started his session. He hadn't eaten or drank anything except water during his exercise, so what was causing the elevated reading? The culprit turned out to be adrenaline. During particularly intense strength training sessions or while performing intervals, our bodies can release large amounts of the hormone adrenaline. When adrenaline is detected by the pancreas, that's a signal that we need more energy. So the pancreas produces a hormone called glucagon, which then triggers the liver into releasing stored sugars. It's those sugars that give us more energy to complete the task.  It's part of the primitive "fight or flight" response in our bodies. When we're going through periods of acute stress, like a particularly tough workout, our bodies want to make sure we have enough energy to survive. The adrenaline starts a chain reaction that ultimately raises our blood sugar for energy. My client was seeing the high glucose readings after his workout BECAUSE his sessions were so intense. In a non-diabetic person, our bodies naturally release insulin to counteract all the glucose. Those elevated blood sugar readings then go down within a couple hours. Unfortunately, type 1 diabetics are unable to produce insulin to handle the spike. If you're a diabetic and engaging in intense activities, you've got a few options. One is to follow any strenuous program with a lower-intensity cooldown, so your body can use up some of the excess blood sugar. But that may only get you part of the way back down. Eat before a workout. For some people, the workout doesn't have to be all that intense for their blood sugar to spike. If you skip breakfast and workout, your body may be pumping out all that sugar because there's no other fuel in the tank. A pre-workout meal gives you energy, the workout will then use up the excess glucose and your body won't have to give an adrenaline kick to keep things going. You should also talk to your doctor about the readings and see if it would be appropriate to take insulin after a workout and how much is recommended. It's important to get the spike under control, you just don't want to take so much insulin it causes a crash later. Exercise is highly beneficial for people with diabetes. It reduces the risk of a heart attack and it makes our bodies more sensitive to insulin. That means when we experience glucose spikes, a fit body can use both natural or injected insulin more effectively to bring things under control. Having a high glucose reading after a workout doesn't mean you must stop exercising, it just means you have to talk to your doctor about the levels. Then talk to a fitness professional about ways to modify your program that will reduce the spikes. That may mean less intense intervals, eating before, or longer cooling down periods afterward. If it's still too high, work with your doctor on a strategy to counter it with an appropriate injection of insulin. Call for a FREE Consultation (305) 296-3434 CAUTION: Check with your doctor before beginning any diet or exercise program.
Nursery rhyme lyrics with its history and origins in English history "A Nursery Rhyme, the first step to your children's education" Old King Cole: lyrics Nursery rhyme lyrics based in history origins dating back to 3rd century Old King Cole ruled part of Britain in the third century. He is reputed to have built the English town of Colchester. In Colchester there is the site of a  Roman gravel pit which is still known today as 'King Cole's Kitchen.'   Cole or "Godebog" was a Dark Age British King, and, a descendant of Britain's pre-Saxon & pre-Roman royal house. The Tudor Kings, starting with Henry VII, claimed to descend from this royal lineage in attempt to further  legitimise the Tudor claim to the English throne. He called for his pipe in the middle of the night And he called for his fiddlers three. Oh there's none so rare as can compare With King Cole and his fiddlers three. Old King Cole nursery rhyme: origins and history
High morning blood sugar common and treatable High morning blood sugar common and treatable By Howard LeWine, M.D. Q: I have type 2 diabetes. My blood sugar runs high in the morning, especially if I am in pain. I have arthritis and neuropathy in my feet and lower legs. Why is my sugar running high in the morning, especially when my pain is worse? A: It’s common for people with type 2 diabetes to have high blood sugar levels in the morning. In type 2 diabetes, formerly known as adult onset diabetes, the body’s cells are resistant to the action of insulin, so sugar doesn’t move readily from the blood stream into cells. Our liver creates blood sugar when we don’t eat to keep our blood sugar from getting too low. This happens in all of us, whether or not we have diabetes. It’s called gluconeogenesis. Normally, insulin controls how much gluconeogenesis happens. The pancreas releases just enough insulin to keep blood sugars in the normal range. If blood sugar starts to rise above normal, more insulin is released. And gluconeogenesis turns off. People with diabetes don’t have this control. With type 1 diabetes, the pancreas does not make insulin. With type 2 diabetes, the problem is insulin resistance. Similar to other cells, the liver cells also have insulin resistance. The pancreas might not release enough insulin to hold back gluconeogenesis, even if the blood sugar level is already high. Pain can also raise your blood sugar. When you are in pain, your body makes extra amounts of stress hormones, such as cortisol and adrenalin. These hormones cause your blood sugar to rise. If you are not taking any insulin now, a small dose of intermediate or long-acting insulin before bed often works well. The insulin will hold back gluconeogenesis during the night and lower your morning blood sugar. If you are already taking insulin, talk with your doctor. He or she can adjust your current doses and the times you take insulin.
Get comfortable because you are about to be astonished by the secrets and appearance of these blue flying wonders. Yes, I said correctly: the blue feathers present in full or partial plumage of some birds are not related to birds’ diet, as it has been proven that the stomach chemicals dissolve the blue pigments contained by fruits, for example (blueberries). Thus, blue pigments do not even reach into the bird’s circulator system to spread in its body and color its feathers or other body parts. Then where does the blue color of birds’ feathers come from? It is not precisely a matter of biology but more likely a matter of physics. The blue color of feathers is related to how the keratin protein disperse the light: having the same (long) wavelength, red and yellow cancel each other when light strikes. Only blue remains reflected, having a shorter wavelength than the first two colors.[1] Further in this article, I invite you to delight yourself with the nine most eye-catching birds with blue feathers. Nine Birds With Blue Feathers Ultramarine Flycatcher Ultramarine Flycatcher Credit: @rajharshanarayanswamy Ultramarine Flycatcher (Ficedula superciliaris) is a small flycatcher (smaller than a sparrow) with different plumage for males and females and a characteristic white stripe above the eyes. The male has an electric-blue plumage on the head, neck, and superior side of its body, and a bright white patch from the middle of the neck, through chest and belly. The female has dark grey-brown instead of electric blue, and the white parts are paler. Ultramarine Flycatcher lives in the mountains and hills forests (pine and broadleaf) from the Himalayas and migrates to southern India in wintertime. As its name suggests, the Ultramarine Flycatcher is an insectivore, the bird foraging insects in the dark spaces from the thick forest. The breeding season lasts from middle April till early July, and the female lays 3-5 eggs in a soft nest made out of moss, fine strips of grass and peeling, hair and roots, usually in holes of trees (about seven meters high), or in a dent on a steep bank. Blue Macaw Blue Macaw Credit: @zafiroytitania Blue Macaw (Anodorhynchus hyacinthinus) is the longest parrot species, the largest flying parrot species, and the largest macaw (1 meter long and between 1.2 and 1.7 kg). Also known as Hyacinth Macaw, Blue Macaw’s plumage is entirely blue, brighter on the upper side, maybe with a shade of grey around the neck to some exemplars, with bright yellow around the eyes and in the laterals of the beak. Blue Macaw fancies the semi-open areas, in general, and it has a variety of favorite habitats: it prefers forests with a dry season that prevents the excessive growth of vegetation, deciduous woods, cerrado and palm plantations, and the palm-Savannas of Pantanal (Brazil), which is like a big luxury oasis for the Blue Macaw. The bird is original from eastern South America (Brazil, Bolivia, Paraguay). Most of the diet of Blue Macaw is made out of palm nuts from two palm species (acuri and bocaiuva palms), and the bird usually harvests the nuts directly from the trees. However, Blue Macaw has another technique to feed itself easier, searching on the cattle lands for excrements of cows containing undigested palm nuts, easier to crack. The bird is supplementing its diet with fruits, nectar, nuts, and seeds. Although Blue Macaw nests even on cliff fronts when constrained by the habitat, it usually makes its nest in tall, old trees (over 60 years old) in an already existing hole, which they enlarge and fill partially with wood bits. The female Blue Macaw lays 1-2 eggs and incubates them for 25-28 days, spending about 70% of its time with the eggs and being fed by the male the entire period. Blue Macaw is a calm and affectionate bird, often called the “gentle giant”, and is very popular as a pet, despite its alarming decreasing number.[2] Asian Fairy-Bluebird Asian Fairy-Bluebird Asian Fairy-Bluebird (Irena puella) is a medium-sized arboreal bird with noticeable variations of plumage colors between males and females. The male Asian Fairy-Bluebird is an exquisite combination of pitch-black and electric- brilliant blue, while the female has a mix of malachite and turquoise green. The Asian Fairy-Bluebird inhabits lowlands forests from southern Asia, the Greater Sudan, and Indochina. Taking them in flight or perching, Asian Fairy-Bluebird feeds mainly on fruits (figs), nectar from flowers, and termites (which also catch from flying), rarely going down on the ground. It is not much known of the breeding behavior of Asian Fairy-Bluebird, but from the captive birds have been reported a mating ritual performed by the male that ends with unison calls, mutual grooming, and even chasing between mates.[3] Siberian Blue Robin Siberian Blue Robin Credit: @birder_falcon Siberian Blue Robin (Larvivora cyane) is a small migratory bird with a notable difference in colors of the plumage between males and females. The male Siberian Blue Robin has electric blue upperparts and white lower parts, separated by black stripe from the eyes till the curve of the wing, while the female is a bit brownish with tinted blue on the upperparts. The Siberian Blue Robin lives in areas with thick, small undergrowth (marshes, taigas, spruce woods) from eastern Asia (Siberia, northern China, Korea, Japan) and migrates in southeastern Asia. As an insectivore bird, the Siberian Blue Robin feeds lurking on the ground after insects and spiders, but it also feeds on grains, seeds, and berries. The Siberian Blue Robin builds a bowl-like nest, padded with grass, leaves, and moss at the base of a tree or other spot with similar protection. The female Siberian Blue Robin lays 2-5 eggs that hatch in about twelve days; both parents care for the feeding and raising of the younglings.[4] Indian Peafowl Indian Peafowl Credit: @nick_peker Indian Peafowl (Pavo cristatus) is a large-sized bird famous for its noble appearance and even more majestic tail. Well, like the most of species we’ve been talking about, the male is the spectacular one, with a blueish-green crown, metallic blue short curly feathers on the top of the head, a bright blue neck and chest, golden feathers on sides and back, a multi-color train with “eyes” on it. The female Indian Peafowl (peahen) also has a crown, the tail is smaller, and the plumage is brown, with shiny green on the neck and white on the lower chest and belly. Native from the Indian subcontinent, Indian Peafowl prefers deciduous forests, margins of forests, and farmlands, near a water supply. However, Indian Peafowl has been introduced in many other countries like the United States, Mexico, Honduras, Colombia, Brazil, Portugal, Croatia, Australia, New Zeeland due to their exquisite beauty. With an omnivorous diet, the Indian Peafowl enjoys a large variety of food: trees or flower buds, petals, seeds, nuts (groundnuts), fruits (Zizyphus berries and even bananas), vegetables (tomatoes, chili), rice, small mammals, reptiles, frogs, and small snakes. The tail’s design, how rich it is, and how many eye dots it has on it is directly connected to the health (immune system especially) of the bird; thus, during the wooing, the male displays its tail to prove to the female that it’s healthy and worthy for breeding. The challenge for the male Indian Peafowl is that it has to spend a lot of time producing and maintaining good tail conditions and investing less time and energy in finding food or defense. The length of the tail and the number of the eyespots can show past health conditions (fights or disease); obviously, the “most wanted” males are those with a rich, long tail and the highest number of eyespots.[5] Blue Rock Thrush Blue Rock Thrush Credit: @mizutama3216 Blue Rock Thrush (Monticola solitarius) is a small to medium-sized bird, with the male having an electric blue head, grey-blue upperparts, brown lower parts, and dark wings. The female Blue Rock Thrush is less strident, with dark brown upperparts and scally pale-brown underparts. Various subspecies of Blue Rock Thrush are widespread around Europe, North Africa, and Southeast Asia (where they are residents); still, some Asian species migrate to sub-Saharan Africa, Southeast Asia, and India in the wintertime. As its name says, the Blue Rock Thrush prefers a rocky habitat in wide-open montane areas, ridges, down of valleys, or (seashore) cliffs. The Blue Rock Thrushes that choose to come closer to human habitats prefer churches and tall buildings, and they can be seen roosting on rooftops or wires. The Blue Rock Thrush is also omnivorous, feeding itself with seeds, berries, worms, insects, spiders, and snails; in some areas, the bird is also known to consume snakes, lizards, and mice. The breeding period of Blue Rock Thrush is approximately from April to July, female building a cup-like nest out of dry vegetation, in a rocky dent, or under a rocky extension. The female lays 3-7 pale blue-green, mottled eggs incubated for around two weeks, and both parents help raise the younglings.[6] Formosan Blue Magpie Formosan Blue Magpie Credit: @dimosb Formosan Blue Magpie (Urocissa caerulea), also known as “long-tailed mountain lady”, is a noisy and strikingly colored bird. Formosan Blue Magpie male and female are similar looking, with black head, a red beak, bright blue plumage, red-brownish wings, white underparts, and long dark-blue tail feathers with white stripes on tips. Formosan Blue Magpie lives in the foothill and lowlands forests, down till gardens, and parks of Taiwan, where it is endemic. With the unusual habit of storing leftovers on the ground (and covering it with leaves) or in the leaves or twigs of the trees, Formosan Blue Magpie is an omnivorous bird and a scavenger. Its diet consists of plants, seeds, fruits, small insects, eggs and chicks of other birds, rodents, snakes, human food waste, and decaying flesh of dead animals (carrion). The breeding season of Formosan Blue Magpie is from March to July, and the monogamous couple builds together a bowl-alike nest in the large branches of trees at the edge of the forest with meadows. The female lays 3-8 eggs, some couples breed the second time right after, and the parents get help from the other birds (younglings from the previous hatches) to feed the baby-birds and defend them. Blue Crowned Pigeon Blue Crowned Pigeon Credit: @tierparkberlin Blue Crowned Pigeon (Goura cristata), also known as Western Crowned Pigeon, is probably the largest pigeon in the world (approximately 66-70 cm in length and 2.1-2.5 kg weight). The Blue Crowned Pigeon’s plumage is blue-grey, with a light-blue crest and a bright white patch surrounded by a reddish-brown patch;  males and females are similar in appearance, except males are bigger. The Blue Crowned Pigeon is endemic to the rain forests of Papua, New Guinea. Living in small groups of two up to ten birds, Blue Crowned Pigeon feeds on the forest floor with seeds, nuts, fruits, and, occasionally, snails. Blue Crowned Pigeon male courts bowing and waving its tail to female and although sometimes it breeds with more than one female, they mate for life. The Blue Crowned Pigeon female usually lays one egg which both parents incubate for about one month. The youngling is fed a mixture called “pigeon milk”, produced by partially digested food. Splendid Fairy Wren Splendid Fairy Wren Credit: @poshbirdphotography Splendid Fairy Wren (Malurus splendens) is a small, long-tailed songbird, with the breeding male looking the most magnificent. The male Splendid Fairy Wren is almost entirely blue, with a black line starting at the beak, continuing near eyes, and joining behind the back of its neck. The female Splendid Fairy Wren is drab brown on the upperparts, with a pale blue tail and reddish beak. The non-breeding male is brownish-grey with a bright blue tail and a little blue on the wings (visible when folded). Splendid Fairy Wren prefers dry and shrubby habitats from semi-arid regions of Australia, where it lives in groups. The bird is also feeding itself in groups, foraging for insects together on the ground or shrubs. The breeding period is between late August and January when the female Splendid Fairy Wren builds an oval dommed nest of dry grass, shreds of bark, spider webs, and roots, with an entrance on one side, close to the soil, usually covered with thorny plants. Also, the female lays 2-4 lays eggs and incubates them alone, but the other birds will help feed the younglings once they hatch. Also Read: Sharing is caring! Similar Posts
Different Types of Athletic Activity Association football, second above soccer, is an enjoyable team sport that also offers many opportunities for social interaction and physical fitness. Sports related to football not only help build up the team spirit but also provide much needed exercise. Numerous sports can result in positive effects on one’s overall health. Many professional sports organizations earn millions of dollars every year from the bets placed on the games. Many non-professional sporting events earn billions of dollars in revenue. Many such sporting events have evolved over time into an industry, with multi-million dollar business empires literally being built on sports gambling. Many major companies now take part in the betting and gaming business. Some of the most well known sports games around include baseball, American football, basketball, softball, hockey, tennis, golf, squash, rugby, Australian rugby and Australian soccer. All of these can be played on a public sports court or even an indoor sports arena. Professional athletics has become a multi-billion dollar industry in the United States. Many colleges and universities have developed athletic departments as a part of the school’s budget. College athletic programs earn billions of dollars in revenue each year. With growing sports interest and concern, sport management and governing bodies have been created to address issues of fairness in sport. These governing bodies are formed by individual sports teams and/or Association Football Associations. Sport management concerns the business aspects of sport related activities including marketing, media relations, performance monitoring, coaching, training and development and prize money. Many government agencies and athletic associations have been created to promote sports competition and to ensure fair play. The NCAA and the Big Ten Conference are two of these associations. Association football has developed its own language and unique culture which is different from that of American football. A major factor is the amount of time and energy devoted to practice, which is called the physical training element. This is often compared to that of chess in that there is more time devoted to tactical exercises during a game of football than to purely tactical strategies. The history of football and the origin of the sport can be traced to Ireland in the 18th century, where they played a variation of the game with a football ball. It was first called Irish football and eventually became known as Association football. Today there are many different spectator sports which are played for entertainment purposes. They are generally considered non-sport or non-sporting but some sports enthusiasts argue that all sports are essentially forms of exercise. The fact that many people enjoy watching football does not mean that they will not use their body in other ways. There is no doubt that all types of athletes engaged in various types of athletic activity and that it enhances their level of health and fitness. Leave a Reply Your email address will not be published.
Piece rate effect on hours worked High piece-rate wages do not reduce hours worked Gregorio Billikopf Encina University of California Effective incentive pay programs mean that workers are paid more--and are more productive--than workers paid on an hourly basis. Employees paid on a piece-rate basis often make higher wages in return for increased productivity than employees paid by the hour. These productivity gains can be substantial. A 1991 study (see California Agriculture September-October 1992), for example, showed labor inputs in vineyard pruning fell sharply from 26 man-hours per acre for employees paid by the hour to 19 man-hours per acre for employees paid by the piece. Most important, increases in worker productivity did not automatically mean lower quality of work. As farmers hear about the potential benefits of incentive pay programs, they sometimes worry that the resulting increase in worker earnings could have a negative effect as well. On days when farm workers are paid on a piece-rate basis, they are often free to leave when they wish. Some farmers have hypothesized that workers have a certain earnings goal for the day. Once this goal is achieved, they reason, workers will go home. Under such conditions, a pay increase may result in workers reducing their hours. Economists would explain this phenomenon as the income effect: increases in income allow those in the work force to take more time for leisure activities. But economists also speak of the substitution effect: the greater the wages, the more a worker has to forfeit by engaging in leisure time. These two forces compete, and other mediating factors, including worker preferences, can also affect results. The main objective of this study was to explore worker feelings and agricultural conditions that may affect a worker's decision to leave early when paid a piece rate. Because wages are influenced by a farm employer's understanding or beliefs about worker behaviors, the results of this study have important policy implications at the farm level for those who make decisions about how to set piece-rate remuneration. If the pay differential between piece-rate and hourly paid work is not enough, farmers may be inadvertently contributing to one of the reasons why incentives sometimes fail to bring out increased worker performance. Worker interviews In the San Joaquín Valley during summer 1993, 510 seasonal farm workers in 15 crews were interviewed. Respondents were surveyed while they engaged in crew work, either hoeing tomatoes; thinning peaches or picking cherries, peaches, grapes, melons or tomatoes. Pay was either by the piece (individual or mini-crew rates), by the hour or some of each. When workers were paid by the hour, they were asked whether they had ever worked on a piece-rate basis. Workers who were working or had ever worked on a piece-rate basis were asked if they generally worked until the end of each day or left early if they had achieved either a production goal (for example, a certain number of boxes harvested) or an earnings goal. Data were also collected on farm worker age, gender and ethnicity and whether they were employed by a farm labor contractor (FLC) or grower. FLCs employed over 70% of the thinning and harvesting crew workers. All but one of the interviews were conducted in Spanish, and all interviewed workers were Hispanic. (An additional employee was Asian and could not be interviewed in English or Spanish.) Few leave early The respondent population included 13% (n = 64) who had never been paid on a piece-rate basis and 87% (n = 448) who had. The following discussion centers on the latter group, with percentages applying only to workers who had been paid on a piece-rate basis. Of the 448 workers who had ever been paid on a piece-rate basis, only a few of the respondents (3%, n = 14) had left work after reaching a wage or production goal for the day. An additional 11% (n = 49) of the respondents also had left early, either from time to time or regularly, but for altogether different reasons. Two key motives offered for leaving early were (1) getting overly hot or tired and (2) not making a sufficient wage (for example, low wages per unit of work performed or not enough to pick). In either case, these workers were generally willing to stay longer if the earning opportunities were greater. Finally, a number of workers explained that infrequently they might leave early to take care of special needs, such as medical visits or other appointments. Not all workers felt free to leave early. Some respondents felt that leaving early could get them fired. A few workers generalized this feeling to every farm operation they had worked for, while others were referring to either a limited number of operations or their present employment. Supervisors of crews paid on a piece-rate basis generally felt workers were free to leave whenever they pleased, even though workers ordinarily stayed until the end of the day. One crew supervisor said that workers frequently left early, but when questioned further explained that this did not happen often in his own crew. An employer who had people leaving early also had a high incidence of turnover, with workers looking for work elsewhere. This FLC did not tell employees how much he was going to pay per tree being thinned until after the work was done, which exemplifies another reason why incentives sometimes fail to bring about increased performance. Finally, leaving early is also a joint decision for those who ride to work with others. Higher paid stay longer The data from this study indicate that those who hire farm workers may generally increase piece-rate wages (or shift from hourly to piece-rate pay) without fear of having workers leave early. In fact, according to the respondents, lower wages were more likely than higher wages to lead workers to an early exit. It appears that the substitution effect plays a greater role in seasonal farm worker motivation than the income effect. Perhaps the most obvious reason for this phenomenon is that farm workers need to maximize their earning opportunities when they can be fully employed. For most crew workers, leisure time is a luxury that they cannot afford to take until a a time of year when there is no work. One might add, then, that the greater the need for income the greater the force of the substitution effect over that of the income effect. That is, increased wage-earning potential among low-income farm workers is not only unlikely to lead workers to seek less work, but actually may motivate them to work even longer hours. Anyone who has observed agricultural workers paid on a piece-rate basis during harvest can attest to the immense effort that farm workers put into work in order to maximize earnings. Farm workers who collect unemployment insurance benefits when off work will also profit from high earnings when they do work. Generally, the greater the worker income received during work periods, the greater the benefits received when unemployed. Some workers and supervisors seemed to differ in their perceptions of whether workers were truly free to leave early. Although these farm employers did not seem to have policies discouraging workers from leaving before quitting time, there are those who do. In today's legal environment, even when paying piece rate, employers must document the total hours worked per day per employee as well as each employee's starting and ending times. Minimum wage and overtime pay requirements can be checked against such documentation. It is no surprise that some farmers prefer to have everyone start and end the workday at the same time. Other findings This study yielded a number of other interesting statistically significant results that deserve some mention. Most of the following observations involve 510 cases. No causal relations are intended. More unmarried respondents worked for FLCs than for growers (46% versus 30%). FLCs were slightly more likely than growers to pay piece rate (77% versus 65%). Eighty-three percent of the males were paid on a piece-rate basis, as compared to 31% of the females. Ninety-two percent of males had been paid on a piece-rate basis at some time, compared to 66% of females. Compared to females, male workers tend to be younger when they enter the agricultural work force and older when they exit. The average farm worker was 30 years old, although the curve was skewed and the mode (the most frequent age) was 22 years. © University of California 2000 Research Papers Directory 15 November 2004
DEV Community Discussion on: Call, Apply, Bind - The Basic Usages alexantoniades profile image Alexander Antoniades Author • Edited on Thank you Sergio! Yes of course..! I don't mention how or when "this" is created. Just stating the fact that every object has a "this" assigned to it! The value of "this" might be anything within the context that is created. Of course you can change that with the methods mentioned in this article ;) sergiokag profile image Sergio Kagiema- Fay Sorry, I did not make it clear. Every object doesn't have a 'this' assigned to it. The 'this' is related to the execution context. And the execution context means how a function is called. That's why I mentioned that a 'this' variable is created when a new execution context is created. :) Thread Thread alexantoniades profile image Alexander Antoniades Author As I've already mentioned in the article....I don't go into much details about "this" and the execution context. The purpose of this article is for a beginner to have a basic understanding of how call, apply and bind works. Saying that every object or function has it's own "this" keyword/variable/object (say it however you want) is enough to make someone understand the basics. Maybe I'll write an advanced article on these concepts next time. thanks again for the feedback Sergio.
Very small liquid or solid particles that are acidic and are small enough to become airborne. Acid aerosol consists of highly rich sulfur particles that leaves the stratosphere of the earth’s atmosphere and have the proportions of a millionth of a meter in diameter. The harmful effects of aerosols are seen in the emission of VOCs that result in smog and other damages throughout the ozone layer.
Hennie Moodie Now, more than ever, attention in South Africa has shifted to how best generate electricity at affordable rates and how to optimise utility management during these uncertain times.  The National Energy Regulator of South Africa (NERSA) recently approved several electricity generation licences for floating power ships, could this become a viable alternative for the future? As the name suggests, a power ship is a special purpose ship on which a power plant is installed that can provide developing countries with ready-to-go infrastructure that plugs into the national grid when required.  Essentially, a power ship is a fully self-contained power station. Each power ship has capacity for its own generation, electrical control, and substation components. The ship also has a maintenance workshop and engineering capabilities. The three licences (awarded to Karpowership SA) are expected to generate 60% of the 2 000 megawatts of electricity targeted under the NERSA Risk Mitigation Independent Power Producer Procurement Programme scheme. And while there have been questions raised in the media about the awarding of these contracts as well as speculative environmental concerns, power ships generate electricity at an affordable, all-inclusive cost, which comprise all capital expenditure, fuel, and all operational and maintenance costs.  The value that these power ships could bring to help alleviate loadshedding concerns cannot be ignored. However, they form only one piece of the increasingly complex puzzle that is the energy crisis in South Africa. Hydrogen potential PwC estimates that South Africa could become an exporter of cost-effective green hydrogen to the world given its immense renewable energy potential.  According to PwC, the infrastructure needed to export hydrogen is similar to existing natural gas networks and is already being piloted in Australia and Japan. South Africa could leverage its existing and new infrastructure to support this initiative. So, despite the potential of power ships, renewable energy remains a key element for dealing with the energy issues the country is facing.  Hardly surprising then that South Africa’s renewable energy sector experienced explosive growth in the past few years and has the potential to create thousands of long-term sustainable jobs and ensure energy security in the future. Asset management But what about the potential of enhancing the efficiency of the national energy grid? Municipalities have already begun implementing their own processes to manage this vital asset as opposed to relying on Eskom to take care of everything.  Being more efficient extends beyond revisiting traditional operational processes – it includes embracing more sophisticated technologies that improve the way in which assets are managed. Putting in place better energy solutions can facilitate more efficient economic growth through a sustainable approach.  New technology is important in this regard, but it can only be truly leveraged if a smart data platform is in place that can provide stakeholders with the required insights to pro-actively manage weak areas in the grid, asset management, and even payment from municipal customers. Overcoming uncertainties A 2018 study identified four major uncertainties that will determine the future of utilities not only in South Africa, but the rest of the continent. Firstly, how will the precarious fiscal position of utilities be resolved to produce long-term sustainability? The solution has the potential to create viable alternative funding models. Secondly, the low quality of governance that currently exists must be addressed through an appropriate regulatory response that facilitates a more enabling environment to use new opportunities in the sector. The third factor to consider as we approach 2022 is the importance of private and public sector partnerships. Seeds have been sown by individual municipalities who are taking charge of their own environments, but there is so much untapped potential waiting to be claimed.  The fourth uncertainty is perhaps the most difficult to plan for as it centres on climate change and its effects throughout the continent. Scenario planning becomes essential. The comprehensive analysis and better understanding of data – whether it is related to power ships or climate change – is key to identifying the most effective ways to optimise the energy sector in South Africa. By Hennie Moodie, New Business Lead at Remote Metering Solutions
Stanford Encyclopedia of Philosophy last substantive content change Saint Thomas Aquinas Life and Works Vita Brevis Thomas was born in 1225 at Roccasecca, a hilltop castle from which the great Benedictine abbey of Montecassino is not quite visible, midway between Rome and Naples. At the age of five, he was entered at Montecassino where his studies began. When the monastery became a battle site -- not for the last time -- Thomas was transferred by his family to the University of Naples. It was here that he came into contact with the "new" Aristotle and with the Order of Preachers or Dominicans, a recently founded mendicant order. He became a Dominican over the protests of his family and eventually went north to study, perhaps first briefly at Paris, then at Cologne with Albert the Great, whose interest in Aristotle strengthened Thomas's own predilections. Returned to Paris, he completed his studies, became a Master and for three years occupied one of the Dominican chairs in the Faculty of Theology. The next ten years were spent in various places in Italy, with the mobile papal court, at various Dominican houses, eventually in Rome. From there he was called back to Paris to confront the hullabaloo variously called Latin Averroism and Heterodox Aristotelianism. After this second three year stint, he was assigned to Naples. In 1274, on his way to the Council of Lyon, he fell ill and died on March 7 in the Cistercian abbey at Fossanova, which is perhaps twenty kilometers from Roccasecca. Philosophy and Theology Many contemporary philosophers are unsure how to read Thomas. As the above sketch makes clear, he was a professional theologian. Nonetheless, we find among his writings works anyone would recognize as philosophical and the dozen commentaries on Aristotle increasingly enjoy the respect and interest of Aristotelian scholars. But his best known work is the Summa theologiae, which is most often cited when Thomas's position on this or that is sought. How can a theological work provide grist for philosophical mills? How did Thomas distinguish between philosophy and theology? Sometimes, Thomas puts the difference this way: "... the believer and the philosopher consider creatures differently. The philosopher considers what belongs to their proper natures, while the believer considers only what is true of creatures insofar as they are related to God, for example, that they are created by God and are subject to him, and the like." (Summa contra gentiles, bk II, chap. 4) Since the philosopher too, according to Thomas, considers things as they relate to God, this statement does not put the difference in a formal light. The first and major formal difference between philosophy and theology is found in their principles, that is, starting points. The assumptions of the philosopher, that to which his discussions and arguments are ultimately driven back, are in the public domain. They are things that everyone knows; they are where disagreement between us must come to an end. These principles are not themselves the products of proof -- which does not of course mean that they are immune to rational analysis and inquiry -- and thus they are said to be known by themselves (per se, as opposed to per alia). This is proportionately true of each of the sciences, where the most common principles just alluded to are in the background and the proper principles or starting points of the particular science function regionally as the common principles do across the whole terrain of thought and being. By contrast, the discourse of the theologian is ultimately driven back to starting points or principles that are held to be true on the basis of faith, that is, the truths that are authoritatively conveyed by the Bible. Some believers reflect on these truths and see other truths implied by them, spell out their interrelations and defend them against the accusation of being nonsense. Theological discourse looks like any other discourse and is, needless to say, governed by the common principles of thought and being, but it is characterized formally by the fact that its arguments and analyses are truth-bearing only for one who accepts Scriptural revelation as true. This provides a formal test for deciding whether a piece of discourse is philosophical or theological. If it relies only on truths anyone can be expected to know about the world, and if it offers to lead to new truths on the basis of such truths, and only on that basis, then it is philosophical discourse. On the other hand, discourse whose cogency -- not formal, but substantive -- depends upon our accepting as true such claims as that there are three persons in one divine nature, that our salvation was effected by the sacrifice of Jesus, that Jesus is one person but two natures, one human, one divine, and the like, is theological discourse. Any appeal to an authoritative scriptural source as the necessary nexus in an argument is thereby other than philosophical discourse. More will be said of this contrast later, but this is the essential difference Thomas recognizes between philosophy and theology. I will conclude this paragraph with a passage in which Thomas summarizes his position. He is confronting an objection to there being any need for theological discourse. Whatever can be the object of inquiry will qualify as a being of one sort or another; but the philosophical disciplines seem to cover every kind of being, indeed there is even a part of it which Aristotle calls theology. So what need is there for discourse beyond philosophical discourse? should be noted that different ways of knowing (ratio cognoscibilis) give us different sciences. The astronomer and the natural philosopher both conclude that the earth is round, but the astronomer does this through a mathematical middle that is abstracted from matter, whereas the natural philosopher considers a middle lodged in matter. Thus there is nothing to prevent another science from treating in the light of divine revelation what the philosophical disciplines treat as knowable in the light of human reason. (Summa theologiae, Ia, q. 1, a., ad 2) Christian Philosophy It will be observed that the formal distinction between philosophical and theological discourse leaves untouched what has often been the mark of one who is at once a believer and a philosopher. It is not simply that he might on one occasion produce an argument that is philosophical and at another time one that is theological; his religious beliefs are clearly not put in escrow but are very much in evidence when he functions as a philosopher. Many of the questions that can be raised philosophically are such that the believer already has answers to them -- from his religious faith. How then can he be thought to be ready to follow the argument whither it listeth, as the objector is unlikely to put it? Furthermore, the inquiries in which the believer who philosophizes engages will often betray his religious beliefs. The importance of this is that a believer runs the risk of accepting bad proofs of the immateriality of thinking and thus of the human soul. On the other hand, a committed materialist may be too quick to accept a bad proof that thinking is just a material process. Such antecedent stances are often the reason why philosophical agreement is so hard to reach. Does it make it impossible? Do such considerations destroy any hope of philosophical objectivity? Surely not, in principle. Believers and non-believers should be able to agree on what counts as a good proof in a given area even if they expect different results from such a proof. Thinking either is or is not merely a physical process and antecedent expectations do not settle the question, however they influence the pursuit of that objective resolution. But the important point is that antecedent dispositions and expectations are the common condition of philosophers, believers and unbelievers alike. Of course, believers hold that they have an advantage here, since the antecedents that influence them are revealed truths, not just hearsay, received opinion, the zeitgeist or prejudice. Thomas and Aristotle As a philosopher, Thomas is emphatically Aristotelian. His interest in and perceptive understanding of the Stagyrite is present from his earliest years and certainly did not await the period toward the end of his life when he wrote his close textual commentaries on Aristotle. When Thomas referred to Aristotle as the Philosopher, he was not merely adopting a facon de parler of the time. He adopted Aristotle's analysis of physical objects, his view of place, time and motion, his proof of the prime mover, his cosmology. He made his own Aristotle's account of sense perception and intellectual knowledge. His moral philosophy is closely based on what he learned from Aristotle and in his commentary on the Metaphysics he provides the most cogent and coherent account of what is going on in those difficult pages. But to acknowledge the primary role of Aristotle in Thomas's philosophy is not to deny other influences. Augustine is a massively important presence. Boethius, Pseudo-Dionysius and Proclus were conduits through which he learned Neo-platonism. There is nothing more obviously Aristotelian about Thomas than his assumption that there is something to be learned from any author, if only mistakes to be avoided. But he adopted many features from non-Aristotelian sources. This has led some to suggest that what is called Thomistic philosophy is an eclectic hodgepodge, not a set of coherent disciplines. Others, struck by the prominence in Thomas of such Platonic notions as participation, have argued that his thought is fundamentally Platonic, not Aristotelian. There is also an undeniable anti-Aristotelian animus in some influential students of Thomas, for example, Etienne Gilson. For Gilson, there is a radically original Thomistic philosophy which cannot be characterized by anything it shares with earlier thinkers, particularly Aristotle. The Order of Philosophical Inquiry Thomas takes "philosophy" to be an umbrella term which covers an ordered set of sciences. Philosophical thinking is characterized by its argumentative structure and a science is taken to be principally the discovery of the properties of kinds of things. But thinking is sometimes theoretical and sometimes practical. The practical use of the mind has as its object the guidance of some activity other than thinking -- choosing in the case of moral action, some product in the case of art. The theoretical use of the mind has truth as its object: it seeks not to change the world but to understand it. Like Aristotle, Thomas holds that there is a plurality of both theoretical and practical sciences. Ethics, economics and politics are the practical sciences, while physics, mathematics and metaphysics are the theoretical sciences. That is one way to lay out the various philosophical disciplines. But there is another that has to do with the appropriate order in which they should be studied. That order of learning is as follows: logic, mathematics, natural philosophy, moral philosophy, metaphysics. The primacy of logic stems from the fact that we have to know what knowledge is so we will recognize that we have met its demands in a particular case. The study of mathematics comes early because little experience of the world is required to master it. But when we turn to knowledge of the physical world, there is an ever increasing dependence upon a wide and deep experience of things. Moral philosophy requires not only experience, but good upbringing and the banking of the passions. Metaphysics or wisdom, is the culminating and defining goal of philosophical inquiry: it is such knowledge as we can achieve of the divine, the first cause of all else. Thomas commented on two logical works of Aristotle: On Interpretation (incomplete) and Posterior Analytics. On mathematics, there are only glancing allusions in Thomas's writings. Thomas describes logic as dealing with "second intentions," that is, with relations which attach to concepts expressive of the natures of existent things, first intentions. This means that logic rides piggy-back on direct knowledge of the world and thus incorporates the view that what is primary in our knowledge is the things of which we first form concepts. Mathematical entities are idealizations made by way of abstraction from our knowledge of sensible things. It is knowledge of sensible things which is primary and thus prior to the "order of learning" the philosophical sciences. This epistemological primacy of knowledge of what we grasp by our senses is the basis for the primacy of the sensible in our language. Language is expressive of knowledge and thus what is first and most easily knowable by us will be what our language first expresses. That is the rule. It is interesting to see its application in the development of the philosophy of nature. Composition of Physical Objects The concern of natural science is of course natural things, physical objects, which may be described as "what come to be as the result of a change and undergo change." The first task of natural philosophy, accordingly, is to define and analyze physical objects. The first thing to notice about this is the assumption that we begin our study of the natural world, not with the presumed ultimate alphabet with which macrocosmic things are spelled, but with a vague and comprehensive concept which encompasses whatever has come to be as the result of a change and undergoes change. The reader of Aquinas becomes familiar with this assumption. Thomas learned it from the beginning of Aristotle's Physics. Now what is to us plain and clear at first is rather confused masses, the elements and principles of which become known to us later by analysis. Thus we must advance from universals to particulars; for it is a whole that is more knowable to sense-perception, and a universal is a kind of whole, comprehending many things within it, like parts. Much the same thing happens in the relation of the name to the formula. A name, e.g. ‘Circle’, means vaguely a sort of whole: its definition analyses this into particulars. Similarly a child begins by calling all men father, and all women mother, but later on distinguishes each of them. (Physics, 1, 1.) Thomas calls the movement from the more to the less general in a science the "order of determination" or specification of the subject matter. The first purchase on natural things is via "physical object" or "natural thing." The "order of demonstration" involves finding the properties of things as known through this general concept. Then, specifying the subject further, one seeks properties of things known through the less common concepts. For example, in plane geometry, one would begin with plane figure and discover what belongs to it as such. Then one would turn to, say, triangle and seek its properties, after which one would go on to scalene and isosceles. So one will, having determined what is true of things insofar as they are physical objects, go on to seek the properties of things which are physical objects of this kind or that, for example, living and non-living bodies. Matter and Form Although natural things are first thought of and analyzed in the most general of terms, there are not of course any general physical objects, only particular ones. Thus, in seeking to discern what is true of anything that has come to be as a result of a change and is subject to change until it ceases to be, Aristotle had to begin with a particular example of change, one so obvious that we would not be distracted by any difficulties in accepting it as such. "A man becomes musical." Someone acquires a skill he did not previously have. Little Imogene learns to play the harmonica. Thomas pores over the analysis Aristotle provides of this instance of change and its product. The change may be expressed in three ways: 1. Man becomes musical. 2. What is not-musical becomes musical. 3. A not-musical man becomes musical. These are three different expressions of the same change and they all exhibit the form A becomes B. But change can also be expressed as From A, B comes to be. Could 1, 2 and 3 be restated in that second form? To say "From the not-musical the musical comes to be" and "From a not-musical man the musical comes to be" seem acceptable alternatives, but "From man musical comes to be" would give us pause. Why? Unlike "A becomes B" the form "From A, B comes to be" suggests that in order for B to emerge, A must cease to be. This grounds the distinction between the grammatical subject of the sentence expressing a change and the subject of the change. The definition of the subject of the change is "that to which the change is attributed and which survives the change." The grammatical subjects of 2 and 3 do not express the subject of the change, only in 1 is the grammatical subject expressive of the subject of the change. Change takes place in the categories of quality, quantity and place, but in all cases the terminology of matter, form and privation comes to be used. The terms bind together similar but different kinds of change -- a subject changing temperature is like a subject changing place or size. Substantial Change The thing to notice about this analysis is that substantial change is spoken of on an analogy with incidental change. The analysis of incidental change is presupposed and regulative. Moreover, the language used to speak of the elements of incidental change are extended to substantial change and altered in meaning so as to avoid equivocation. Thomas sees Aristotle's philosophical vocabulary arising out of analysis of what is most obvious to us and then progressively extended to more and more things insofar as the later is made known by appeal to the prior. Thus we can see that matter and form apply in a graded and connected way to the various kinds of incidental change and then to substantial change. It is this feature of Aristotle's language that Thomas adopts as his own. It both provides the lens through which Aristotle can be correctly seen and it provides a rule for knowing and naming that will characterize Thomas's use of Latin in philosophy and in theology as well. Perception and Thought When the discussion moves on from what may be said of all physical objects as such to an inquiry into living physical things, the analyses build upon those already completed. Thus, "soul" will be defined as the substantial form of living bodies. The peculiar activities of living things will be grouped under headings like nutrition and growth, sense perception and knowing and willing. Since a living thing sometimes manifests an instance of such activities and sometimes does not, they relate to it in the manner of the incidental forms of any physical object. And this provides Thomas, as he feels it had Aristotle, with a path of procedure. Let us skip to the cognitive activities of living things. How can we best analyze perception -- that is, seeing, feeling, hearing, and the like? In continuity with what has gone before, the questions are put in this form: How best to analyze coming to see, coming to feel, coming to hear, and the like. Seeing these on the analogy of change as already analyzed, we look for a subject, a privation and a form. The sensing subject is, say, the animal, but the proximate subjects to which they are attributed are the powers of sight, touch, hearing, and the like. An instance of seeing is describable as the power's moving from not seeing to seeing. Since the object of seeing is color, the change from not seeing to seeing issues in the power having the form of color. We could give as an example of physical change, a substance acquiring a color. Now, while there are physical changes involved in sensation -- the organs are altered in the way physical bodies are -- that is not the change involved in perception as such. That is, in feeling a body my hand's own temperature is altered by the contact. But feeling cannot be just that, since any two physical bodies that come into contact would undergo a similar alteration of temperature. Feeling the temperature, becoming aware of it, is another sort of change, however much it depends on a contemporary physical change in the organs of sense. Having the color or temperature in this further sense is thus made known and named by reference to physical change. The fundamental difference between the two ways of acquiring a form is this. In a physical change of color, the change produces a new numerical instance of the color. In grasping or sensing a color, a numerically new instance of color does not result. We have here the basis for talk of immateriality in perception. If the acquiring of a form by matter in physical change results in a new instance of the form and this is not the case with perception, we can make this point that acquiring the form in sensation is not identical to the acquiring of the form by matter in the primary sense. Thus, we both want to speak of the subject of sensation on an analogy with physical change and to distinguish the former from the latter. This is done by speaking of the immaterial reception of a form. Nonetheless, the sense power is a subject and thus matter in an analogous sense. In his interpretation of Aristotle's De anima Thomas defends a view that was as contested in his own time as it is almost an orphan in our own. Among the tenets of so-called Latin Averroism was the view, first held by Averroes, that the move from perceptive acts to intellection is not one from a lower to a higher set of capacities or faculties of the human soul. When Aristotle contrasts intellection with perception and argues that the former does not employ a sense organ because it displays none of the characteristics of perception which does employ an organ, he is not, on the Latin-Averroistic view, referring to another capacity of the human soul, the intellect, but rather referring to a separate entity thanks to whose action human beings engage in what we call thinking. But the cause of this, the agent intellect, is not a faculty of the soul. (Aristotle distinguished two intellects, a passive and an active.) The proof for immortality which results from a wholly immaterial activity is therefore a statement about the incorruptibility of the separate entity, not a basis for arguing that each human soul is immortal because it has the capacity to perform immaterial activities. The Latin-Averroists consequently denied that Aristotle taught personal immortality. Given this consequence, Thomas's adoption of the opposite interpretation -- viz. that the agent intellect is, like the passive intellect, a faculty of the human soul -- may seem merely an interested desire to enlist Aristotle's support for a position in harmony with Christian belief. Thomas is frequently said to have baptized Aristotle, which seems to mean that he fitted him to the Procrustean bed of Christian doctrine. Of course, the full Christian view is not simply that the soul survives death but that it will be reunited with body, and Thomas nowhere suggests that there is any intimation of this in Aristotle. Oddly enough, it is often friends of St. Thomas who suggest that he used Aristotle and was not chiefly concerned with what Aristotle might actually have intended. Surely this is libelous. It would be less of an accusation to say that Thomas got a passage wrong than that he pretended it meant something he knew it did not. But the important point, all these centuries later, is whether Thomas's reading is or is not supported by the text. When he commented on the De anima, he seems not to be concerned with the flare up in Paris over Latin Averroism. This is the basis for dating the commentary in 1268, before Thomas returned to Paris. The commentary, accordingly, cannot be read as though it were prompted by the controversy. Of course, Thomas might be said to have long term interests in taming Aristotle to behave in a Christian way. As it happens, during the second Parisian period, in the thick of the Latin-Averroist controversy, Thomas wrote an opusculum dedicated to the question: what did Aristotle actually teach? The work is called in the Latin, De unitate intellectus contra averroistas. I have translated it as, On there being only one intellect. This little work is absolutely essential for assessing the nature of Thomas's Aristotelianism. He provides us with an extended textual analysis to show that the rival interpretation cannot be sustained by the text and that the only coherent reading of the De anima must view the agent and passive intellects as faculties of the human soul. His interpretation may be right or wrong, but the matter must be decided on the basis of textual interpretation, not vague remarks about Thomas's intentions. Beyond Physics Jaeger found both these conceptions of metaphysics clumsily juxtaposed in a crucial passage of Book Six. One might indeed raise the question whether first philosophy is universal, or deals with one genus, i.e. some one kind of being; for not even the mathematical sciences are all alike in this respect, -- geometry and astronomy deal with a certain particular kind of thing, while universal mathematics applies alike to all. We answer that if there is no substance other than those which are formed by nature, natural science will be the first science; but if there is an immovable substance, the science of this must be prior and must be first philosophy, and universal in this way, because it is first. And it will belong to this to consider being qua being -- both what it is and the attributes which belong to it qua being. (1025a24-33) Jaeger's hypothesis, which cannot withstand half an hour's scrutiny, dominated interpretations of the Metaphysics until yesterday. Giovanni Reale's book had to await English translation before it could have any impact. By that time, people were turning from Jaeger and toward Aristotle neat, but this was only to weary of Jaeger, not to disprove him. Thomas's reading of the Metaphysics makes clear how mistaken Jaeger's claims are. Mathematical things, on the analogy of ‘concave’, do not have sensible matter in their definitions. Lines, points, numbers, triangles -- these do not have sensible qualities whether stated universally or singularly. The fact that we define mathematicals without sensible matter does not commit us to the view that mathematicals actually exist apart from sensible matter. What then is the subject of metaphysics? "Subject" here means the subject of the conclusion of the demonstrative syllogism. The discussion of definition in effect bore on the middle terms of demonstrative syllogisms and the suggestion is that formally different modes of defining, with respect to removal from matter and motion, ground the formal difference between types of theoretical science. The subject of a demonstration in natural philosophy is defined without singular but with common or universal sensible matter; the subject of a mathematical demonstration is defined without any sensible matter. How can the subject of metaphysics be expressed? The possibility of the science depends on our knowing that some things exist apart from matter and motion. Mathematics does not presuppose the separate existence of its objects; metaphysics does. Why not then say that metaphysics deals with things separated from matter and motion, that is, as Jaeger notes, with a particular kind of being? But that is the not the subject ever assigned to this effort by Aristotle. The methodological reasons can be found in chapter 17 of Book Seven of the Metaphysics: the subject of a science must always be a complex entity. That is why the subject of this discipline is being as being. Philosophical and Scriptural Theology Philosophical theology is not some science distinct from metaphysics; it is simply the name that can be given to metaphysics because it appeals to God as the cause of its subject. This may make it seem that knowledge of God is merely a bonus, a tangential consideration; on the contrary, it is the chief aim of the science. But the divine can only be known indirectly, through its effects. For this reason, metaphysics can be viewed as an extended effort to examine substance in order to come to knowledge of the first cause. And given the principle that we name things as we know then, this can be regarded as a prolonged effort to devise a language with which to speak of God. Analogous Names Aristotle spoke of "things said in many ways", a notable instance of which is "being." One of the difficulties with assigning being or being as being as the subject of a science is that a subject must be univocally common to the things that fall under it and being is not univocal but has a plurality of meanings. Aristotle solved this problem with his account of "things said in many ways," by observing that while they have many meanings, these form an ordered set and one of the meanings is primary and regulative. Substance is being in the primary sense and that is why the science of being as being is effectively a science of substance. Thomas's term for such names is analogy: ‘being’ is an analogous term and its primary analogate is substance. In the crucial middle books of the Metaphysics -- Seven and Eight -- we have an analysis of substance which takes off from material substance, which is a compound of matter and form, and arrives at a notion of substance as form alone. This definition does not fit material substance, of course, but it is devised in order to be able to apply the term substance to the immaterial things whose existence has been established in natural philosophy. This extension of names whose natural habitat is sensible things to God is another instance of analogous naming for Aquinas. Names common to God and creature bring out another feature of our knowing insisted upon by both Aristotle and Aquinas. If we ask what the primary analogate of names common to God and creature is, the answer is: the meaning of the term as it applies to creatures. The word must be refined before it can be applied to God and this means the formation of an extended meaning which leans on the primary meaning for its intelligibility. Essence and Existence It is evident that material substances exist contingently. They come into being and they pass out of being and while they exist, existence is not what they are. Thomas accepts from Boethius that it is self-evident that what a thing is and for it to exist differ (diversum est esse et id quod est). Material things depend upon causes to exist, both to become and to be. There is no need to dwell on this except insofar as it provides a springboard to speak of immaterial substance. Only in God is it the case that what he is and that he is are identical: God is existence. The phrase Thomas uses to express this is ipsum esse subsistens. Of course this is paradoxical. Existence is the actuality of a substance, not itself something subsistent. This is true with material substances, but when we ask what we mean by saying that God exists, we have to negate aspects of material existence in order to avoid speaking of Him as if he were a contingent being. The problem that Thomas now faces is how to speak of the immaterial substances which are less than God although superior to material substances, that is, angels. For a material thing to exist is for its form actually to inhere in its matter. But what is it for a pure form to exist? Since immaterial substances less than God are dependent on the divine causality in order to exist, existence cannot be what they are, of their essence. In short, in angels too there is a distinction of essence and existence. Thomas notes that a created separate substance is what it is and not another thing: that is, it has the perfection it has, but not unlimited perfection. It is a being of a kind, not being as such. Gabriel is perfect as to Gabrielitas, but he is not Raphael or Michael. Form thus operates as a restriction on existence as such. In God alone is there unrestricted existence; he is existence, ipsum esse subsistens. Moral Doctrine When Aristotle sought to isolate the human good, he employed the so-called function argument. If one know what a carpenter is or does he has the criteria for recognizing a good carpenter. So too with bank-tellers, golfers, brain surgeons and locksmiths. If then man as such has a function, we will have a basis for deciding whether someone is a good human being. But what could this function be? Just as we do not appraise carpenters on the basis of their golf game or golfers on the basis of their being able to pick locks, we will not want to appraise the human agent on an incidental basis. So too we do not appraise the carpenter in terms of his weight, the condition of his lungs or his taste buds. No more would we appraise a human being on the basis of activities he shares with non-humans. The activity that sets the human agent apart from all others is rational activity. The human agent acts knowingly and willingly. If this is the human function, the human being who performs it well will be a good person. Many have come to this point, pulse quickened by the possibilities of the function-argument, only to be gripped with doubt at this final application of it. Rational activity seems too unmanageable a description to permit a function-analysis of it. Of course Aristotle agrees, having made the point himself. Rational activity is said in many ways or, as Thomas would put it, it is an analogous term. It covers an ordered set of instances. There is the activity of reason as such, there is the activity of reason in its directive or practical capacity, and there are bodily movements and the like which are rational insofar as they are directed by reason. If the virtue of a function is to perform it well, the analogy of "rational activity" makes clear that there is a plurality of virtues. Moral virtues are habits of appetite brought about by the direction of reason. Temperance is to seek pleasure rationally, courage is to react to the threat of harm rationally. The virtues of practical intellect are art and prudence; the virtues of theoretical intellect are insight, science and wisdom. All this and much more enters into Thomas's moral teaching. Here as elsewhere the Aristotelian component looms large. Thomas will distinguish acts of a man from human acts, the former being activities truly found in human agents but also found in other non-human agents too. The human act is one which proceeds from knowledge and will. Since the human act by definition is the pursuit of a known good, the question arises as to the relationship between the objects of the myriad acts that human perform. Is there some over-all good sought by human agents? Is there an ultimate end of human action? In commenting on chapter two of Book One of the Nicomachean Ethics where Aristotle argues for there being an ultimate end, Thomas points out that the argument is actually a series of reductiones ad absurdum. That is, the denial of an ultimate end of human action reduces to the claim that there is no end to human seeking, that it is pointless. This analysis has not gotten the attention it deserves: the implication is that it is self-evident that there is an ultimate end which is why denials of it must flounder in incoherence. The argument for ultimate end that Thomas puts forth in the Summa theologiae is somewhat different. Any action aims at some good. A particular good by definition shares in and is not identical with goodness itself. What binds together all the acts that humans perform is the overarching goodness they seek in this, that and the other thing. That over arching goodness, what Thomas calls the ratio bonitatis, is the ultimate end. It follows that anything a human agent does is done for the sake of the ultimate end. Natural Law What is natural law? One description of it is: the peculiarly human participation in the eternal law, in providence. All creatures are ordered to an end, have natures whose fulfillment is what it is because of those natures. It is not peculiar to man that he is fashioned so as to find his good in the fulfillment of his nature. That is true of anything. But others things are ordered to ends of which they themselves are not conscious. It is peculiar to man that he becomes aware of the good and freely directs himself to it. Of course man is not free to choose the good -- any choice is a choice of the good. But as to what is really as opposed to only apparently his good, he is not free to make that what it is. He is free to direct himself or not to his true end, however. He says that our first conception is of being, of that which is, and our first judgment is that you cannot affirm and deny the same thing in the same sense simultaneously. Since few if any humans first utter ‘being’ or its equivalent and no one fashions as his first enunciation the principle of contradiction, facts as known to Thomas as ourselves, his meaning must be more subtle. It is this. Whatever concept one first forms and expresses verbally -- Mama, hot, whatever -- is a specification or an instance of that which is. Aristotle has observed that children at first call all men father and all women mother. The terms then function as generic for any male or female. Even more basically, each presupposes that what is generically grasped is an instance of being. Being is prior not because it is grasped absolutely, without reference to this being or that. It is some particular being that is first of all grasped and however it is named it will mean minimally something that is. Most moral judgments are true, if true, only by and large. They express means to achieve our overall good but because there is not a necessary connection between the means and end, they can hold only for the most part. Thus there are innumerable ways in which men lead their lives in keeping with the ultimate end. Not all means are necessarily related to the end. Moral philosophy reposes on natural law precepts as common presuppositions, but its advice will be true only in the main. It might be noted that when Thomas, following Aristotle, says that man is by nature a social or political animal, he does not mean that each of us has a tendency to enter into social contracts or the like. The natural in this sense is what is not chosen, but given, and what is given about human life is that we are born into the community of the family and are dependent on it for years in order to survive. The moral consists in behaving well in this given setting. Thomas's teaching came under attack, largely by Franciscans, immediately after his death. Dominicans responded. This had the effect of making Dominicans Thomists and Franciscans non-Thomists -- Bonaventurians, Scotists, Ockhamists. The Jesuits were founded after the Reformation and they tended to be Thomists, often with a Suarezian twist. Vatican II, the ecumenical council that met from 1962-1965 spelled the end of the Thomistic Revival. It was widely held that the Council had dethroned Thomas in favor of unnamed contemporary philosophers. (When they were named, quarrels began.) In the post-conciliar period, Catholics have managed to board every sinking ship in sight and now with the vogue of the notion that modernity has failed and the Enlightenment Project come a cropper, many, Catholics and others, are turning to Thomas as a spur or foil for their thinking. In 1998 John Paul II issued an encyclical called Fides et Ratio which may be regarded as the charter of the Thomism of the third millennium. Life and Works Other Internet Resources Related Entries Maritain, Jacques Copyright © 1999 Ralph McInerny Stanford Encyclopedia of Philosophy
Chaos in The Currency Markets : Currency Crisis of The EMS 1. What does the crisis of September 1992 tell you about the relative abilities of currency markets and national governments to influence exchange rates? The currency markets and national governments both have abilities to influence exchange rates. Like other financial markets, foreign exchange markets react to any news that may have a future effect. Speculators are the part of the currency markets that take currency positions based on anticipated interest rate movements in various countries. Day-to-day speculation on future exchange rate movements is commonly driven by signals of future interest rate movements. By using the signal, speculators usually take the position before the things actually occurred. Sometime, if high power enough, the speculators position can influence the exchange rate movement. The government controls is one of the factors affecting exchange rate. The government can influence the equilibrium exchange rate in many way, including direct intervening (buying and selling currencies) in the foreign exchange markets and indirect intervening by affecting macro variables such as interest rates. 2. What does the crisis of September 1992 tell you about the weakness of fixed exchange rate regimes? From European currency crisis of September 1992, it shows us that there are weakness of the fixed exchange rate system. When exchange rate are tied, a high interest rate in one country has a strong influence on interest rates in the other countries. Funds will flow to the country with a more attractive interest rate, which reduces the supply of fund in the other countries and places upward pressure on their interest rates. The flow of fund would continue until the interest rate differential has been eliminated or reduced. This process would not necessarily apply to countries outside ERM that do not in the fixed exchange rate system, because the exchange rate risk may discourage the flow of funds to the countries with relatively high interest rate. However, since the ERM requires central banks to maintain the exchange rates between currencies within specified boundaries, investors moving funds among the participating European countries are less concerned about exchange rate risk. 3. Assess the impact of the events of September 1992 on the EU "s ability to establish a common currency by 1999. A major concern of a common currency is based on the concept of a single European monetary policy. Each country"s government may prefer to implement its own monetary policy. It would have to adapt to a system in which it had only partial input to the European monetary policy that would be implemented in all European countries, including its own. The system would be alike to that used in the U.S. , where there is a single currency across states. Just as the monetary policy in the U.S. cannot be separated across different states, European monetary policy with a single European currency could not be separated across European countries. While country governments may disagree on the ideal monetary policy to enhance their local economies, they would all have to agree on a single European monetary policy. Any given policy used in a particular period may enhance some countries and adversely affect others. There are some other concerns that could prevent the implementation of a single currency. For example, at what exchange rate would all currencies be cash in to be exchanged for the common currency to be used? (think about the trouble after reunification of Germany). It would be difficult to reach agreement on this question for each European country"s home currency. Also, some economists believe that changing exchange rates serve as a stabilizer for international trade. Thus, the lack of an exchange rate mechanism could possibly cause greater trade imbalances between countries. 4. The crisis of September 1992 occurred because the ERM system was too inflexible. Discuss. The inflexible system was not the main reason. The main reason is because there are too different monetary policies among the member of ERM. The German government was more concern about inflation and less concerned about unemployment because its economy was relatively strong. On the other hand, other European governments were more concerned about stimulating their economies to reduce their high unemployment levels. This argument was proved at the end of the crisis when Germany and France "s government joined forces to defend the franc against speculative pressure. If all the member joined forces early the crisis may not occur. 5. If you were an executive for a company that engages in substantial intra-EU trade, how would you react to the events of September 1992? Because the company engages in substantial intra-EU trade, the exchange rate risk is not the major issue-under fixed exchange rate system the exchange rate will fluctuate narrowly. A major concern is the interest rate movement. High interest rate results in high cost of capital to the company and slow growth economic. The problem will even more serious if the company have to pay floated rate liabilities in foreign currencies. The company should consider hedging against interest rate risk such as using interest rate swap or using fixed rate liabilities.
More than you wanted to know about Ad Blockers (updated 2021) Updated on 10.01.2022 What is an ad blocker? Ad blockers are used to block or remove advertising from websites. They work by removing the JavaScript files that are responsible for the ads on a website. Ad blockers come in two forms: browser extensions and mobile apps. Browser extensions work by blocking ads before they appear on your screen, while mobile apps filter content directly at the source, so you never see an advertisement at all. Some people install ad blockers to improve their browsing experience because they do not want to be bombarded with advertisements while they are trying to read articles or watch videos online. Others might use it for security reasons, as some advertisers have been known to serve malware through advertisements. When were ad blockers invented? The first ad blocker was invented in 2003 by a programmer named Wladimir Palant. He had a problem with pop-up ads and decided to use his knowledge of programming to create a fix for this problem. Wladimir Palant created AdBlock Plus based on a initial version created a by than student from Copehangen. Henrik Aasted Sørensen figured out that the best way to prevent loading of ads would be to block any requests towards the handful of adservers at the time. How many people use ad blockers? Ad blockers are on the rise, and some countries are approaching 50%. There will always be people who use adblockers; they simply don't want their internet experience to be cluttered with ads. This inevitably leads to less revenue for publishers. Ad blocking is not just about limiting ads. It can also be used to limit tracking cookies and other unwanted content. It can even help to avoid malware downloads that are found in ads. Clearly, there are many advantages to using ad blockers! Is it ethical to use an ad blocker? A lot of people think that the use of an ad blocker is unethical because it is a form of stealing content from the website. However, others believe that an ad blocker is a form of protest against internet advertising and not taking away from the site content. How do Ad blockers make money? Recently ad blockers have found a new way to make money. They use their technology and expertise to offer website owners advice on how they can make more money from advertising through other channels, such as affiliate marketing or sponsored content, or by offering subscriptions and paywalls for content. It has been discovered that consumers who use ad blockers spend more money on other products and services than users who do not use an ad blocker. Companies are realizing this and offering a premium for displaying their ads. Do ad blockers work on Youtube? Ad blockers are assumed to be able to block all ads on Youtube. However, this is not the case. Ad blocking is used by many people as a way to reduce the amount of ads they see and it has been proven to work on most sites and apps. However, YouTube is an exception. Ad blockers can't block Youtube ads because Google uses a different form of advertisement technology called DFP (DoubleClick for Publishers) which does not require ad servers. The reason why ad blockers don't work on YouTube is that Google relies on DoubleClick for Publishers, which doesn’t require third-party ad servers like those that most other publishers use. Do ad blockers work on Facebook? Facebook is one of the most heavily used sites in the world and it generates a lot of revenue from ads and sponsored posts. Ad blockers do work for Facebook. However, with the rise of ad blockers, Facebook has started to find other ways to combat the practice. It's practically a game of whack-a-mole between users using AdBlockers and Facebook software engineers. What types of ad blocking technologies are there? The term "ad blocker" can represent any kind of content blocking technology that blocks ads on websites. Pi-hole is a network-level ad blocker with DNS level filtering, which blocks most ads your computer sees. It also takes care of advertisements on mobile devices and blocking trackers. A chrome extension is an ad blocker that operates in the browser and intercepts requests to advert servers. VPNs are another popular form of ad blockers as they work by altering the user's IP address so their device loads from a different country which has less strict regulations on advertising. AdBlock Plus review AdBlock Plus is one of the most popular ad blocking tools available today, and with good reason. It has been around for over 6 years now, and continues to maintain its position as one of the leading free ad blockers available. The software can be installed on any browser you desire, including Chrome, Safaribook, Firefox, Opera Mini (yes it’s still around!), Internet Explorer (you’re welcome Microsoft), and even Safari. You can also use AdBlock Plus on your Android devices if you’ve got an AdBlock Plus-optimized version of your browser. uBlock review uBlock is another popular ad blocking tool, which is also free to use. The software boasts over 1 million downloads on the Chrome store alone, with approximately 7 million downloads running on all browsers. The extension is compatible with most browsers, including Chrome, Firefox, Safari, Edge, and Opera. Poper blocker (for Chrome) review Poper Blocker is a Chrome extension that helps in keeping a tab on the problematic content in a web page. It blocks all pop-ups and other intrusive advertisements from the web pages, which helps in better content consumption. More AdBlockers will be reviewed shortly
Canadian Hydroelectric Transmission Line Will Bring In More Renewable Energy Than All Minnesota Solar Projects – But None of It is “Renewable” Last week, Minnesota Power announced the completion of its Great Northern transmission line, which will import 1.4 million megawatt hours of hydroelectric power from Manitoba, Canada, to Northern Minnesota, enough to power nearly 150,000 Minnesota homes for an entire year. Interestingly, this single transmission line will provide 34 percent more electricity than all of Minnesota’s solar panels, combined. However, because of the way Minnesota’s renewable energy mandate is written exactly zero of this renewable electricity will be considered “renewable” by the Minnesota state government. Solar Is Less Productive Than Hydro Power According to the U.S. Energy Information Administration, Minnesota generated about 1 million megawatt hours of electricity with solar in 2018, or enough to power about 110,500 households. While this may not sound the alarm bells for the casual reader, it definitely should, because it shows how much more productive hydroelectric power is than solar. The Great Northern Transmission line will have a capacity to transmit 250 megawatts of power, which could provide about 2.1 million megawatts hours, or enough to power 232,000, if the line were operating at full capacity. This means Minnesota Power would be utilizing more than 60 percent of the line’s capacity. One the other hand, Minnesota had 734 megawatts of solar installed, but generated just a little over 1 million megawatt hours in 2018. If solar had been generating at full capacity, it would have generated about 6.5 million MWh, meaning Minnesota’s solar displays produced only 16 percent of their potential output. In energy lingo, this percentage of output is called a capacity factor. The low productivity of solar means we spend a lot of money for only a little energy. If we take the National Renewable Energy Laboratory (NREL) cost estimates for solar and apply them to Minnesota’s solar installations, we can estimate that Minnesota has spent about $1.5 billion on installing solar panels from 2016 to 2018. In contrast, Minnesota Power paid just $325 million for the GNT, although it will pay an additional cost for the electricity it purchases, and these costs do not appear to be publicly available. Even though hydroelectric power is obviously renewable, exactly zero of the electricity purchased from Canada will count toward satisfying Minnesota’s renewable energy mandate, the Next Generation Energy Act (NGEA). This is because the NGEA specifically excluded large hydroelectric generators to promote the development of wind and solar. Some people have argued that Minnesota shouldn’t count the hydroelectric power from Canada because one of the goals of the NGEA was to build renewable power plants in Minnesota and provide local jobs. However, many Minnesota utilities are building wind facilities in North and South Dakota to meet their mandated renewable percentages, which is not creating local jobs. If an energy resource is renewable, it should be counted as renewable. The double standard at play here is arbitrary and decidedly non-scientific. It has also likely increased the cost of electricity paid by Minnesota families and businesses and resulted in higher carbon dioxide emissions. In other words, Minnesota’s liberal policies have resulted in our state having the worst of both worlds, higher electricity prices and higher emissions than if we had embraced hydroelectric power as a renewable resource. Higher Electricity Prices As all of our regular readers know by now, Minnesota’s electricity prices have increased 30 percent faster than the national average since 2005, when the legislature mandated Xcel Energy to put wind on its system, and the NGEA made things worse in 2007. This has forced Minnesota families to pay more to keep their lights on. It has also forced energy intensive businesses in Minnesota, like farming, manufacturing, and mining to pay more to keep their doors open. In some cases, high energy prices are a deal breaker, and these businesses leave our state, in others, it simply makes Minnesota businesses to become less competitive with their competition in other states, and countries. Conservatives should be outraged that Minnesota’s early experiment with the Green New Deal has produced such bad economic outcomes. Minnesota liberals, especially those who claim climate change is an “existential crisis” like Governor Tim Walz, should be outraged that emission are likely higher due to our reliance on wind and solar than they would be if we had invested early in hydroelectric power. Higher Carbon Dioxide Emissions Hydroelectric power is much more reliable than wind or solar because the fact that water his held behind a dam and is released to generate electricity when the energy is needed acts as a natural battery. This means hydroelectric power, unlike wind or solar, can be stored until it is needed, as you can see in the diagram below. This makes hydro electric power far more useful than wind or solar, because it is self-storing, and it can be used to generate electricity whenever you need it. This dispatchability means that unlike wind and solar, which always require backup power plants when the weather isn’t cooperating, hydroelectric power can more easily be a one-to-one replacement for other power plants like coal or natural gas plants. This means they are also more able to replace the carbon dioxide emissions that accompany these fuels. In fact, Minnesota could easily be 100 percent renewable if it were to embrace, rather than exclude, hydroelectric power from Canada. Manitoba Hydro has a current generating capacity of 5,690 megawatts, which could power enough to power up to 5.2 million Minnesota homes. In other news, Minnesota only has 2.2 million homes. Manitoba Hydro could also add more dams to their system and produce more energy. The company will complete a 695 MW facility in 2021, and the company has the potential to build 4,300 MW of additional capacity. Together, these power plants would be able to power 9.9 million Minnesota homes for an entire year. Political Barriers to Hydroelectric Power Hydroelectric power appears to be a no-brainer, so why haven’t Minnesota politicians legalized large hydro facilities for the NGEA? I’ll let you, dear reader, fill in that blank. Conservative lawmakers have tried to change the NGEA multiple times to allow large hydro to count toward our renewable energy mandates, but it has been systematically opposed by liberal politicians. Their opposition is rooted in the fear that allowing hydro to count toward the mandate would replace the need for wind and solar on the electric grid. In essence, they are afraid that a level playing field for renewables would allow a superior product to win out over their preferred pet projects. This is ultimately why it is impossible to take advocates of a Minnesota Green New Deal seriously. If they really care about the environment, they should be seeking solutions that have a maximum environmental return on investment. Instead, they are stubbornly forcing Minnesotans to have the worst of both worlds.
Exhibition about colours at the GRIMMWELT Kassel Foto: © GRIMMWELT / N. Klinger Foto: © GRIMMWELT / N. Klinger Why is Little Red Riding Hood’s cap red? What associations does Snow White’s hair awaken? And how do we actually find a blue beard? In fairy tales, colours do not only make the stories more colourful, they also have cultural, historical and symbolic meaning. In the new special exhibition “Red Hood, Blue Beard” at the GRIMMWELT Kassel, which Archimedes designed, visitors can immerse themselves in the world of colours in fairy tales.  It can be visited from 13 November 2019 to 13 April 2020. By means of textual examples and contemporary illustrations visitors embark on a journey through the complex meaning and symbolism of colours. In addition, games and experiments await the visitors, as the era of Romanticism was not only the Golden Age of fairy tales but also the scientific exploration of colours, conducted for example by Philipp Otto Runge or Johann Wolfgang von Goethe. Interactive installations developed by Archimedes bring optical phenomena to life and offer impressive photographic motifs. They play with our visual habits and encourage participation. Visitors can create colourful shadows, continue their own reflection to infinity, stand under Mother Hulda’s golden rain and let off steam with a round of “colour dots”. Further Information: https://www.grimmwelt.de/en/exhibitions/red-riding-hood-blue-beard/ Even more news Updates from our projects, research, and exhibitons
Heat rash: How long does prickly heat take to go away? HEAT RASH is a rash usually caused by excessive sweating and is common around heatwaves. How long does prickly heat take to go away? Comment section Share on LinkedIn Share on Pinterest Copy link Link copied This Morning: Alison Hammond shares 'Heatwave Hacks' in 2019 Heat rashes or prickly heat is more likely to occur in hot weather. It’s uncomfortable, unsightly, and itchy, so you’ll want to try and avoid heat rash. It is normally harmless and is easy to prevent. Heat rash is normally caused by blocked sweat glands. When it is hot and you sweat excessively, sometimes your sweat glands can become blocked. This trapped sweat can cause a heat rash or prickly heat to occur a few days later. Heat rash is common in babies who cannot control their body temperature, but is also seen in adults and children. READ MORE- How to sleep through a thunderstorm Heat rash: Woman itching heat rash Heat rash: Heat rash is really common, but that doesn't make it easy to deal with (Image: Getty) Heat rash symptoms The symptoms of heat rash are the same in adults and children, and it can appear anywhere on the body. You may find that the rash will spread to different areas of your body, but it isn’t infectious to others. The symptoms are: • small red spots • an itchy, prickly feeling • redness and mild swelling Heat rash: Heat rash on arm Heat rash: Heat rash should go away within a few days (Image: Getty) Heat rash: Heat rash on neck Heat rash: Heat rash is caused by blocked sweat glands (Image: Getty) How to avoid heat rash The best thing to do to avoid heat rash in hot weather is keep cool and don’t sweat. To do this, you should wear loose cotton clothing in light colours. You should use sheets at night instead of duvets and sleep in cotton pyjamas. Drink lots of water– at least six cups of water a day– to avoid becoming dehydrated. Heat stroke symptoms: How to know if you have heat exhaustion [INFORMER] Heat rash: Three ways to soothe symptoms [INSIGHT] Is your fan blowing out hot air? How to make your fan COOLER [EXPLAINER] How to calm heat rash Heat rash is itchy and uncomfortable, so you will want to treat it as soon as you can. A pharmacist might recommend calamine lotion, antihistamine tablets, or hydrocortisone cream. Apply something cooling such as a damp cloth or ice pack wrapped in a tea towel for 20 minutes maximum. If your skin is really itching, try to pat it instead of scratching it. Avoid using strong perfumed shower gels or creams, instead using baby products and products for gentle skin. Heat rash: Sweat rash Heat rash: Sweat rash looks very similar to heat rash (Image: Getty) How long does prickly heat take to go away? Heat rash should clear up on its own after a few days. If it doesn’t, you could be suffering from something more severe. If this is the case, ring your GP and ask for advice. You could be suffering from the jock itch, which is a common infection in the groin or armpit area caused by staying in sweaty clothes after exercising. The rash could actually be seasonal eczema, which is triggered by pollen, sweat and suncream. Another condition you could have instead of heat rash is sweat rash. This is common in overweight people who gather sweat in their fat creases and folds.
A Covid testing kit The Omicron strain has been described as ‘an unprecedented sampling’ of mutations from four earlier variants of concern: Alpha, Beta, Gamma and Delta © FT montage/Dreamstime Just weeks after scientists in South Africa and Botswana first alerted the world to the emergence of a worrying new coronavirus variant, researchers are beginning to understand the implication of its mutations, which far exceed the number on any previous variant. Omicron’s 50 genetic changes include more than 30 on the spike protein, the exposed part of the virus that binds with human cells. Scientists expect these changes to make it more transmissible than the dominant Delta variant and more likely to evade the immune protection provided by vaccines or previous infection. Why are scientists alarmed about Omicron? Initial fears were fuelled by Omicron’s highly unusual genetic profile. Jeffrey Barrett, director of the Covid-19 Genomics Initiative at the Wellcome Sanger Institute, described it as “an unprecedented sampling” of mutations from four earlier variants of concern: Alpha, Beta, Gamma and Delta, together with other changes never seen before whose significance is still unclear. Fifteen of the mutations are on the receptor binding domain that acts like a “grappling hook” for the virus to enter human cells, said Jacob Glanville, a computational immunologist and founder of Centivax, a US therapeutics company. Graphic showing the key mutations that shape the Omicron variant Such extensive mutations make it harder for the human immune system, trained by vaccination or previous infection to recognise a different variant, to tackle Omicron. Current vaccines are based on the original strain first detected in the Chinese city of Wuhan two years ago. Some of the mutations also increase transmissibility regardless of immunity. “From the perspective of evolution, Omicron is the most dramatic evidence of natural selection I’ve ever seen in any organism,” said Barrett. How fast is Omicron spreading? Though Omicron was first detected in samples from southern Africa, no one knows for sure where it originated and how it accumulated so many changes. South Africa was the first country to report a fourth wave of Covid driven by Omicron, starting in Gauteng province, which includes the cities of Johannesburg and Pretoria, and then spreading nationwide. But cases in Gauteng seem to have peaked last week and are now declining. Elsewhere, Omicron cases are multiplying fast across the world, with numbers doubling within as little as two days in some places and governments responding with measures to restrict social mixing. The UK and Denmark have reported the fastest Omicron growth in Europe, though this is partly a result of their having more thorough genomic surveillance systems than other countries. The UK Health Security Agency on Monday reported 8,044 confirmed and 25,534 probable Omicron cases, bringing the confirmed and probable total above 170,000. Fourteen people with Omicron infection have died and 129 have been hospitalised in the UK. How well will Covid vaccines work against Omicron? Two lines of evidence are confirming virologists’ fears that vaccines and prior infection will be less effective at preventing cases of Omicron than previous variants. One comes from lab experiments in which scientists expose blood samples to the virus and measure the antibody response. The other uses “real world” epidemiology to estimate how vaccination status affects the risk of developing Covid. Early conclusions suggest that a large decline in immunity against Omicron occurs after two vaccine doses, particularly among people double-jabbed with the Oxford-AstraZeneca product, while a booster jab works well at restoring most, though not all, of the protection. Chart illustrating that English data show substantially increased risk of breakthrough infections with Omicron, but boosters push risk back down towards Delta levels On Monday, Moderna said a booster shot of its mRNA vaccine increased antibody levels 37-fold if a half dose was injected and 83-fold for a full dose. BioNTech and Pfizer had previously said a third dose of their vaccine achieved a 25-fold antibody boost, though the studies are not directly comparable. But lab studies focus on neutralising antibodies rather than the longer-term immunity given by T-cells, which can “remember” past infections and kill pathogens if they reappear. And the Omicron wave has not been going long enough for epidemiologists reliably to estimate efficacy against hospitalisation and death, which is expected to be much higher than against mild breakthrough infections. How does the disease caused by Omicron compare with previous variants? Early signs from South Africa have been encouraging, as people infected with Omicron seemed to show milder symptoms than those in Delta patients. Many were asymptomatic. Others suffered from coughs, fatigue and aches. Chart showing that preliminary data from South Africa suggest that among adults, the share of Covid patients requiring acute or intensive care is lower than in past waves However, scientists warn that it is still hard to make direct comparisons between Omicron and previous variants, not only because of the time lag before severe symptoms appear but also because coronavirus immunity in the population through previous infection and vaccination is different today. Epidemiologists at Imperial College London on Friday published a study of cases in England which, they said, “finds no evidence of Omicron having lower severity than Delta, judged by either the proportion of people testing positive who report symptoms, or by the proportion of cases seeking hospital care after infection”. However, they added, “hospitalisation data remains very limited at this time”. Laboratory experiments are beginning to illustrate Omicron’s distinctive pattern of activity across different human cell types, which affects disease symptoms. Studies at the universities of Hong Kong and Cambridge suggest that it is significantly less efficient than other coronavirus strains at infecting lung cells and tissue, while it propagates much faster in bronchi higher up the respiratory system. These differences, if confirmed, would help to explain Omicron’s greater ability to pass between people while apparently causing less severe Covid disease. “Even with the strong booster campaign, Omicron has to be very significantly less severe than Delta to not likely overwhelm healthcare capacity levels,” said Ewan Birney, deputy director-general of the European Molecular Biology Laboratory. “Data from South Africa is interesting and positive about some lower severity of Omicron, but not obviously clear cut that it is at the levels needed.” Get alerts on Coronavirus pandemic when a new story is published Copyright The Financial Times Limited 2022. All rights reserved.
Know About Low Impact Development The work of stormwater engineers included capturing rainwater and moving the water away as fast as they can. This process led to the growth of a vast system of large, centralized therapy facilities handling billions of gallons of stormwater annually – a very efficient, but not very environmentally-minded solution. Green Stormwater Infrastructure (GSI) is one of the subsets of low impact development (LID) which is a method that focuses on stormwater management with methods that emulate and/or enhance the natural cycle of water. You can get the construction of to get low impact development. Image source google After years of development that emphasized efficiency over ecology and the short-term convenience over long-term advantages Recently, there has seen a shift in the way we think about the design challenge. GSI seeks to reduce how much stormwater is that is transported to these huge municipal infrastructures and reduce the impact of overflows and flooding that can result from their failures. Stormwater engineers trade pipe flow acceleration equations with soil infiltration accelerations and create systems that hold rainwater off the area in the area it falls, returning the rainwater to the natural ecosystem when it is cleaned. The result of this shift is that the majority of land developers and owners are aware of the benefits of adopting green stormwater infrastructure strategies. The efficiency of these systems is in their design, installation, and maintenance but also the effective education of the people affected by these systems. Developers and owners of sites must be aware of the importance of these systems, to assist the engineers and architects who develop them. Tooth Savior – Root Canal Treatment When people learn about dentistry, especially root canal treatment, they are generally overcome with dread and terror. Although dental treatment may sound uncomfortable, it has the potential to save your teeth. This procedure isn't nearly as uncomfortable as it once was, thanks to advances in current technology. You will experience less pain using bore surface needles if you obtain dental care from a trained and experienced dentist. To learn more about tooth canal treatment, read the entire article. To know more about Root Canal Treatment visit What is a Root Canal? A root canal is a space inside the root of a tooth. It is a naturally existing space within a tooth that is made up of the pulp chamber, the main canal, and more complex anatomical branches that can hold the root canal to the root surface. Each root may have one or two canals. The dental pulp is the tissue that forms the dentin portion of the tooth. In this treatment, the infected area is cleaned until it is free from bacteria. After disinfecting the affected area, it is disinfected and hermetically sealed, and later sealed with a crown. How is the treatment done? Treatment is needed when the root of the tooth becomes infected or enlarged due to injury or advanced caries. Anesthesia is applied to the tooth and the dentist will drill a hole through the crown of the tooth to access the pulp chamber. Special files are used to pull the infected pulp out of the canal. The canal is filled with a permanent material like gutta-percha, which helps to keep the canals infection-free. This treatment is a way to save the tooth and soothe the terrible pain. Usually, healing is achieved by placing a cap on the affected tooth. As everyone knows this treatment is very painful. But think of treatment this way: You have to go through some degree of pain to get rid of the pain. So in this case, a root canal can be a tooth saver. Signs You Need a New Muffler In Greece Those of us unfamiliar with cars can still tell when something needs fixing. If we hear strange noises and the car doesn't sit "properly", it may be time to head to the garage. When it comes to your muffler, there are a few ways to let you know it's time to replace it. How can you tell if your car force exhaust needs to be replaced? Pay attention to the following signs when driving. 1) Pay attention to gas consumption. It's always a good idea to calculate the mileage from the last tank purchased each time you refuel. If you find yourself burning gas faster than usual and driving no further, it's time to have your exhaust checked.  2) Watch out for a smoke. If one day you smell something unusual in your cars, such as an unpleasant odour or an odour that you would associate with a broken-down car, you should have it checked immediately. Problems with your exhaust can help increase smoke emissions. 3) Watch for changes in your car when you start the engine. With a good exhaust, your car engine will lift and purr well. If the sound is louder and less subdued, you may need to have your exhaust checked. While all of these clues may indicate the need for a new exhaust, they can also point to other problems with your car. For example, your exhaust may be fine, but other parts of the car may be responsible for the anomaly. Whatever the case, you don't want to risk car trouble. Waiting too long can result in a repair that costs more to repair, or you may lose your warranty. Features of the Customized Boxes How Long Should You Sit In A Massage Chair Almost everybody who uses massage chairs doesn’t know how long should they be sitting in it. They like the vibrations and warmth so much that they keep seated in the massage chair. So, we are going to discuss this particular topic today that how long should you sit in a massage chair. When we talk about the recommended time, then it is 15 minutes, but people who are in excruciating pain tend to sit for a lot longer than that. But those who have just started to use a massage chair should stick to a 15-minute timeframe. Spending too much time on a massage chair can bruise your muscles and cause inflammation in the soft tissues of the body. Most massage chairs feature an automatic timer that lets people know not to go over a certain time limit. Some units have massage programs that shut off the unit when used for too long. Why Start With Short Massage Increments? You can stay longer on the massager chair as you become comfortable with it. You need to understand that being on a massage chair is like a workout, so the initial days can give you a different feel, but as you begin to use it, your body will get accustomed to it. So, just like any other product, start slow and then, increase your time on the massage chair and you will get the maximum benefits from it. Choose Golf Shirts That Are Made From The Right Materials The way you dress in the golf shirts you wear may not be as significant as golf balls, or golf clubs you use when playing in golf. But it can say something about your personality or appearance. It makes its own unique impact on the game. It makes you feel confident and happy. You present your image to others as someone who is taking his game seriously and should not be dismissed lightly. However, inappropriate attire could be seen as reflecting your image and could negatively impact your concentration and eventually the game. In addition, wearing inappropriate attire can cause you to feel uncomfortable when playing.  To prevent unwanted distractions, the following guidelines can help you pick the appropriate ladies golf shirts that will ensure you are comfortable at the golf course. ladies golf shirts Even with the strict rules, of course, there are still plenty of golf shirts to select from. Pick classic and straightforward designs like knit jumpers and Polo shirts. Don't wear shirts with vulgar or offensive messages or shirts sporting large and massive logos or colorful colors.  A casual outfit will not be permitted as well, so dress in the right golf shirt that is stylish. Get your game going by choosing the correct material for the golf shirts you'll wear. Highly advised are polo shirts made of natural fiber that is woven in a manner that air flows freely across the material.  It is important to have shirts with varying thicknesses to let you adjust to the changing weather conditions. Thin and lightweight clothes are perfect for summer, while sweaters with a thicker knit and long-sleeved shirts are great in winter. Wear layers to easily adjust to changing temperatures and feel comfortable during your game. How to Pick Apt And Affordable Eyeshadow Palette A makeup kit would not be complete without a stunning eyeshadow palette. There are many cosmetic companies offering eyeshadow palettes in a wide variety of shades, so selecting the perfect eyeshadow for your collection is not a daunting job. Picking the right eyeshadow colors for perfect Looks. While choosing an eyeshadow palette is simple, there are some aspects to be taken into consideration to ensure you get the most effective. If you want to have perfect eye makeup then you can buy the best quality pretty eyeshadow palettes online. 17 Best Eyeshadow Palettes 2021 - Eyeshadow Palettes for Every Eye Color Image Source: Google Be sure to purchase palettes from reputable brands since these eyeshadow palettes aren't just more durable, but also offer greater quality and variety of shades.  If you enjoy playing with different colors, then you should consider more of a palette that has a variety of shades. Simple shades are ideal for those only interested in just a handful of shades.  In addition to the appropriate palette, it is also necessary to have high-end eyeshadow brushes to ensure the perfect use of the eye shadow. How to Identify Your Skin Tone for The Perfect Eyeshadow Color Match If you are a brunette skin tone, try to select pearlescent tones. To get a stylish and chic look you can choose from charcoal and navy blues, or deep plums. Golden bronze shades compliment more dark skin tones and are an excellent shade for the daytime.  There are a lot of online websites that allow you to obtain an exclusive selection of eyeshadow palettes featuring a wide range of stunning shades.  The Importance of Learning Statistics If you're an undergraduate student studying statistics, then you've probably experienced situations where you're stuck with a dilemma at an unorthodox time of the day and there is no one around to help.  We are in the Internet age, where everything is possible. All you have to do is connect to your personal computer, and be linked to the Internet. There are numerous websites created specifically to help you with statistics online. These websites provide a forum for both experts and students studying statistics to exchange ideas and resolve questions. You can also look for online statistics tutor in Brisbane The Importance of Statistics in Many Different Fields - Udemy Blog Image Source: Google There are numerous websites offering statistics assistance for no cost while others charge a minimal amount. You can also get help with statistics from tutors that are private to your region. There are websites that offer private tutoring across a variety of subjects including biology, math, English, physics, for instance.  Finding a tutor who is private in your neighborhood will free you from having to travel and aid in getting a thorough knowledge of the matter. If you're not confident regarding the basic concepts of statistics, and probability, this could cause difficulties in solving a challenge successfully. If, for instance, you do not understand the concept of probability you'll find it difficult to solve a challenge that is related to probability.  Understanding the concepts well is crucial to be able to tackle the issues. The majority of the time, students who are slow learners have it difficult to comprehend the concepts being taught in class. Because there are many students in the same class, it's impossible for a teacher to focus on every student individually. Understanding Generalized Anxiety Disorder (GAD) GAD or Generalized Anxiety Disorder is characterized by too much unrealistic anxiety about the aspects of life such as social relationships, financial matters, work issues, etc.  Some people develop GAD as a child and may not see symptoms until they reach adulthood. Either way, living with GAD may last a long time and in many cases, occurs with other anxiety or mood disorders. How common is GAD? There are about four million Americans who suffer from chronic generalized anxiety disorder symptoms. GAD often begins during childhood or adolescence but may also start in adulthood. Women are more likely to develop GAD compared to men. How is GAD diagnosed? When symptoms of GAD are present, healthcare providers will evaluate by asking questions about one’s medical and psychiatric history; your physician may also require you to undergo a physical examination.  Healthcare providers base the diagnoses on reports of the intensity of the symptoms which also includes any problems that are related to functioning caused by the symptoms. Doctors determine the degree of dysfunction and indicate a specific anxiety disorder.  Symptoms for GAD may include: • Constant worrying or obsessing about a particular concern that is either big or small. • Not being able to let go of worry. • Restlessness, and feeling on edge all the time • Excessive worry • Fear of making the wrong decisions or getting distressed about decision making • Exaggerated anxiety about everyday life for no reason. A person with GAD always tends to think that disaster is going to happen anytime soon. • Physical signs and symptoms: • Fatigue • Muscle tension or muscle aches • Irritability • Trembling, feeling twitchy • Easily startled • Difficulty in sleeping • Nausea, diarrhea, irritable bowel syndrome, or IBS • Headaches Key Considerations To Choose Soft Serve Mix All ice cream consists mainly of milk, flavor, sugar, water, and air. Regular ice cream contains about 30 to 60% air. Soft serve ice cream contains more air, usually about 70% by volume.  The extra air gives the soft serve a softer texture and makes it melt faster, hence the name. You can make a soft serve in a homemade ice cream maker with a soft-serve mix. You can also buy the best soft serve mix in Australia through various online sources. Important consideration when choosing a soft-serve mix Powder vs. liquid soft-serve mixes If you are used to buying soft serve ice cream in restaurants and fast-food restaurants, then you know best about liquid soft serve ice cream. The liquid and creamy soft serve mix is based on milk and has a higher fat content and fewer ice crystals than the powder mix. No mixing is required before pouring the liquid mixture into the soft-serve machine. Liquid soft serve mixes are usually sold in bags rather than in boxes, which also saves storage space. Powder soft-serve mixes can last longer on the shelf. They are usually dairy-free, which makes them fat-free and less expensive than liquid mixes. To make a smooth powder mixture, the powder must be mixed with water.  Non-dairy products You can choose a dairy-free mix for a soft serve for a number of reasons. The soft powder mixes inherently exclude dairy, and dairy-free blends are effortless to find for sorbets or other frozen fruit-based desserts. Proteins and fats can thicken a soft-serve mixture to get the creamy taste of a soft-serve without the use of dairy products. Many dairy-free soft blends use soy, coconut oil, or pea protein for their softness.
How the Rugged F4F Wildcat Held the Line During World War II Designed by Grumman Iron Works, the tough little fighter kept America in the fight during those early dark days in the Pacific Theater F4F Wildcat The F4F Wildcat, a carrier-based aircraft, saw service from the surprise attack at Pearl Harbor in 1941 to the Japanese surrender aboard the USS Missouri nearly four years later. NASM As Joe Foss scanned the skies on January 25, 1943, he spotted a large formation of 100 Japanese bombers and escort planes flying toward Guadalcanal. With only 12 fighters in his attack group, the Marine Corps captain recognized he was impossibly outnumbered. That didn’t stop Foss, though. The cigar-chomping, tough-as-nails American ace with 26 kills in World War II gunned his Grumman F4F Wildcat into action and ordered his pilots to use the overcast conditions to their advantage as they attacked the overwhelming enemy force. By darting in and out of the clouds, the aptly named Foss’ Flying Circus quickly shot down four Japanese fighters and tricked the bombers into thinking they had encountered a much larger squadron. The Japanese withdrew from attacking strategically important Henderson Field on Guadalcanal and flew back to their base without dropping a single bomb. Foss was hailed for his heroics by saving the airbase with a “brilliant tactical maneuver” and would receive the Congressional Medal of Honor, which put him on the cover of Life Magazine. By equaling Capt. Eddie Rickenbacker’s First World War record of planes destroyed, Foss became America’s first “ace of aces.” Though his total would later be surpassed by other U.S. fighter pilots, Foss’ exploits came at a time when the outcome of the war was very much in doubt and the country desperately needed heroes. “His remarkable flying skill, inspiring leadership and indomitable fighting spirit were distinctive factors in the defense of strategic American positions on Guadalcanal,” his citation read. Foss answered the call of duty piloting the F4F Wildcat, a very capable carrier-based aircraft that saw service from the surprise attack at Pearl Harbor in 1941 to the Japanese surrender aboard the USS Missouri nearly four years later. While not as fast or nimble as other planes, it was surprisingly rugged and effective as a fighter in the hands of a trained pilot. “The Wildcat was actually well-armored and very difficult for the Japanese to shoot down,” says Larry Burke, curator of U.S. naval aviation at the Smithsonian's National Air and Space Museum, which has a Wildcat in its collections that will go on view in the museum's new World War II gallery slated to open in 2024. “They had to put a lot of fire into it to knock it out. The Wildcat and other Grumman planes during the war acquired this reputation for sturdiness. One of the reasons the company was nicknamed Grumman Iron Works was because of the survivability of the planes,” says Burke. Joe Foss In his Grumman F4F Wildcat, the tough-as-nails American ace Joe Foss lead his squadron to a victory in the Pacific Theater in January 1943. Wikimedia Commons And rugged it was. Despite the quickness and agility of the F4F’s main opponent in the Pacific Theater—the Japanese Mitsubishi A6M Zero—the Wildcat shot down an average of seven enemy aircraft for every one of its losses. Foss scored his first kill in an F4F during his first combat mission on October 13, 1942. However, his Wildcat was also severely shot up and was being chased by three Zeros. He was forced to land at Henderson Field at full speed with a dead engine and no flaps. Nearly a month later, on November 7, another Wildcat flown by Foss was hit hard, forcing him to ditch at sea. He survived both incidents with no major injuries. In fact, Foss was shot down at least four more times and got back in the cockpit after each one. Bigger and better fighters—such as the Grumman F6F Hellcat—would enter the fray, but the F4F Wildcat saw service throughout the war. Its durability, along with multiple .50-caliber machineguns and wing space for two 100-pound bombs, made it ideal for search-and-destroy missions. Plus, it possessed another feature that worked wonders onboard the smaller escort carriers often used in convoys. “(Company co-founder) Leroy Grumman very famously experimented with an eraser and unfolded paperclip to come up with a diagonal hinge that he called the Sto-Wing,” Burke says. “It allowed the wing to fold alongside the aircraft, kind of like a bird’s wing, which permitted more planes to be stored on the escort carriers. “What’s fascinating is the wing is still in use today,” says Burke. The E-2 Hawkeye, the Navy’s early-warning aircraft, features the Sto-Wing because the aircraft has a big radome on top that interferes with the upward-folding wings on other aircraft. Wildcat fighter aircraft wing Leroy Grumman very famously experimented with an eraser and unfolded paperclip to come up with a diagonal hinge that he called the Sto-Wing,” says Smithsonian curator Larry Burke. Mark Avino, NASM The Wildcat at the National Air and Space Museum is a version of the Grumman F4F-4, known as an Eastern Division FM-1. It was actually produced under license by General Motors, which began manufacturing Wildcats in 1943 when Grumman switched over to F6F Hellcats. While this plane never saw combat, it is a well-preserved example of the aircraft that helped turn the tide in the Pacific. “The Wildcat was the airplane that held the line for the Navy at the beginning of World War II,” Burke says. “It held on in combat until better aircraft could be designed and put into service. It was the primary fighter operated by the U.S. Navy and Marine Corps. “The Wildcat was the main fighter used in four of the first five big carrier battles of World War II. And it continued to make a mark later in the war by flying from the convoy escort carriers, which the bigger planes couldn’t do,” says Burke. During World War II, Joe Foss scored all 26 of his victories flying the F4F, including his last three kills in a single day. He would go on to serve at age 39 as the youngest governor in South Dakota history, first commissioner of the upstart American Football League in 1959 and was a passionate gun-rights advocate as president of the National Rifle Association from 1988 to 1990. Foss died in 2003 at the age of 87. Wherever he went later in life, the Wildcat was never far from his heart. The Medal of Honor fighter pilot developed a strong affinity for the rugged aircraft that protected him and propelled him to victory at a critical moment in history. “I loved that Grumman F-4F Wildcat,” Foss once told an interviewer. “I just loved it. It was like a brother to me. It was tough, and I would be able to destroy the enemy with it, and I couldn’t blame it for the speed it didn’t have, but I felt at home.”
Things About Software You Need To Experience It Yourself. Software application is a series of guidelines which inform a computer exactly how to perform a certain operation. For example, software program which informs a computer system to switch on a details home appliance, or software which informs a computer system to execute an on the internet transaction. Both examples entail a specific item of hardware. Nonetheless, software is usually kept inside a computer system. The computer which saves the software more than likely has a circuit card or mommy board which acts as a database for the software application. A significant difference between hardware and software is that software serves a low-level operation job while equipment offers a higher degree or a user-level operation. As an example, allow’s consider just how an automobile drives. The cars and truck drives, the engine transforms power right into a mechanical movement, and tires offer traction. In this instance, we can see just how software program serves a reduced level operation job while hardware serves a higher level or user-level operation. Software application, nevertheless, is made to execute a greater level job. As well as to do so, it must connect with particular equipment components. So for instance, allow’s take a look at the next example. When an individual inserts a bank card right into a hardware tool, claim a credit card equipment, the machine does what is called a “big salami” procedure. This suggests that the computer system requires to review the details that is on the debit side of the card and afterwards refine the purchase (giving it a “cost”). Software is usually less expensive than hardware because it doesn’t require to support a big variety of various features. For instance, allow’s take software application like the Windows os and also compare it to a program like Java. Windows functions simply great if you are just thinking about fundamental functions. Java on the other hand will run efficiently if the program you are running has a large selection of different features and uses up a great deal of resources (a Java applet) when it is not proactively being used. Software application like Java is a lot more pricey to establish because it likewise needs to have a large collection of various different kinds of Java code which can be run during the runtime of an application. Software like Windows is more affordable to develop because there are fewer commonness in between different pieces of hardware and also the os. Software program can also be much less expensive due to the fact that it does not have to consist of device drivers which are actually needed to operate a particular item of computer. Many software comes preinstalled with devices like printers and also key-boards. Windows comes preinstalled with all of the standard features like computer mouse, key-board, screen capture tool, electronic camera, video capture device etc. That’s why the command line trigger, which is primarily a series of really basic commands to do something, is always included as part of Windows. As well as the motorist is often consisted of with the os at the time of the installment of the hardware. Therefore the first thing that you require to be aware of is the difference between energy software application and also software. Utility software application aids you utilize the basic operating system features and also supplies you with a variety of typical uses for the equipment that is present in your computer. As an example word processing software and also workplace efficiency software, are both utility software program. On the other hand software has different types of commands which you can carry out on the computer. They can be command line instances, which are simply basic text commands to do something, to produce a data, or to print something. An additional example is shell commands, which are executed by the command shell. These examples are not so usual but are necessary for the procedure of certain programs. Energy software application is made to be really basic to make use of and also to be able to perform a details collection of jobs. However utility type applications are not the just one that you will certainly locate on a computer system. Other kinds of applications are system software and application software. In a feeling system software is required even if you don’t intend to utilize any type of type of energy application. Yet if you wish to utilize some energy type programs, you can make use of such applications such as Disk Cleaning Software which aids you to clean up your hard disk. Software is a series of guidelines which tell a certain computer how to perform a details task. In comparison to hardware where the maker is developed and also essentially does the work, software program actually executes the desired work and also is constructed by the user. Basically, software application are made use of to transform how a computer system works, as well as the new software is then installed or downloaded and install. There are various kinds of software, each created for a particular function. A lot of computer system systems use some kind of software for their procedure. One of the most popular is the Windows operating system. The reason Windows is so prominent is since it is what most individuals know as “PC”. Equipment based running systems differ in that they run straight from hardware without requiring to be plugged into a PC. Both of these kinds of operating systems have various objectives, nonetheless. For instance, in Windows, all the documents, applications, as well as other alternatives are organized in a tree structure. Each documents or program alternative is connected to a branch, and afterwards the next branch down is the choice which was clicked. When a designer or someone in advertising and marketing wants to alter exactly how a piece of software program jobs, they will likely require to experience this whole tree system to get software development solutions. That being said, it may be extra efficient to get software program advancement solutions from the designers themselves, rather than having to go with the entire Windows system. This way, designers can focus on coding and also less on the technicalities of the Windows operating system. freight tracking software Designers also make use of data source administration to make the computer system as reliable as possible. The data source management system makes it feasible to have multiple versions of a particular application, or multiple variations of a program, going for the exact same time. Database management additionally aids with software application design by making the making of technical remedies easier. Database engineering is comprised of database style, data evaluation, database optimization, as well as assimilation with the rest of the engineering team. An effective database management team has the skills to resolve technological troubles while making use of one of the most efficient shows languages and ideal database readily available. Leave a Reply
Virtual reality software is occasionally used to treat chronic pain, depression, and anxiety. A program that promotes an idyllic setting may distract a user from feeling discomfort or a sense of unease. An individual's sense of well-being may heighten and their heart rate or blood pressure may be lowered to a healthier level than what was initially experienced. An Overview Of A VR Session A medical provider may prescribe their patient to use a VR software program during a hospital stay or while a patient is at home. A virtual reality headset and a display screen are operated simultaneously. While a patient is wearing a headset, they will view a video through the eyewear apparatus that is connected to a headset. A patient's movements will coincide with the footage that they are supplied with. Sounds that correspond to the setting will be heard through an audio device that is built into a headset. Panoramic views will be featured, which can make a virtual reality session seem authentic. Forests, lakes, waterfalls, and mountains are some geographical features that are often used during VR sessions. These types of features may prompt a patient to feel at peace during a session. Telehealth Management Sessions Telehealth management sessions are conducted while a client is at home. VR supplies are furnished by a medical provider. A client will be instructed on how to set up the equipment. A VR session can be conducted while sitting in a comfortable chair or while reclining on a couch or a bed. A medical provider will observe a VR session remotely. They will view and listen to the same VR recording as the individual who is being treated. Throughout a VR session, a client can verbalize how they are feeling. Specific parts of a viewing experience may be more helpful in blocking pain than others. For instance, if a client prefers the sound of crashing waves and tends to feel relaxed when viewing a body of water from a particular angle, a medical provider will take note of this and may adjust future VR programs to focus more on the sights and sounds that a client responds to in a positive manner. Some software programs may require the use of a heart monitoring device. Medical equipment like this will allow a clinician to observe when a patient's heart rate drops. Adjustments to a viewing experience may be made based upon the calculations that are collected. Contact a clinic that offers at-home VR therapy for pain for more info.
Multi-dimensional Spectroscopy: A brief intro Some days I call myself a spectroscopist—a scientist whose primary tool is light. When I think of light, I think of it as an electromagnetic wave—I’ll call it an electric field from now on. In lab I use electric fields to tickle samples. But I don’t just use one field, I use many fields. In my lab these fields are laser beams. I would like to give you a sweeping overview of what my lab does. To do this, I first need to give you a bit of perspective/context. I expect this post to merely whet your appetite. In the future I plan to write about more specialized topics which have to do with my research (like your request of explaining detection strategies: homodyne vs. heterodyne, or how you can do frequency or time domain measurements). In days of old the predominant type of spectroscopy was vanilla absorption/reflection/transmission spectroscopy. Light from a source, like a tungsten lamp, is shone on a sample. The amount of light both incident on the sample and reflected off of (transmitted through) the sample is quantified. A mathematical metric is then used to define the reflectance (transmittance) spectrum. Different colors of light interact differently with the same sample. So, a spectrum of intensity vs. incident color can give information about a sample. In general chemistry labs which I have taught, students use absorption spectroscopy to find the concentration of metal ions in solutions, and monitor the progression of a reaction. It is fun to note that the human eye works by observing the light transmitted and reflected off of surfaces and then spectrally resolving that light to give us information about our surroundings. In other words, the human eye/brain uses vanilla absorption/reflection/transmission spectroscopy to interrogate the world. My lab specializes in multi-dimensional spectroscopy. This is in contrast to 1-dimensional spectroscopy like the spectroscopy outlined in the previous paragraph. There are many ways for a spectroscopy to colloquially be called multi-dimensional, but the most common definition requires the spectroscopy to entail the usage of at least two different colors of electric fields. These interrogating fields can be scanned in their color. Information is then represented as 2D plots of color_1 vs. color_2 with some measured intensity as a z-axis which is generally represented as a heat map. There is another way for a spectroscopy to be multi-dimensional. You can have multiple electric fields which are pulses in time (they can be the same color). You can then change the amount of time present between each pulse. Information is then represented as 2D plots of color vs. delay_time with some measured intensity as a z-axis which is generally represented as a heat map. My lab uses both of these types tricks to do our spectroscopies. Alexander, I presume you are currently furious with me! I have introduced complexity without telling you why. Why should I go through the effort of having multiple colors of laser pulses and then change their time delays? The traditional answer: because the world is so complex and interesting that vanilla 1D spectroscopy can’t begin to tell us everything about a sample. A more nuanced answer divides the rational into multiple categories. Some people do multi-dimensional spectroscopy because the extra dimensionality allows them to understand the ground state[footnote: By the ground state, I mean the unexcited and not perturbed system] of complex systems which have 1D spectra which are impossible to interpret due to “spectral congestion”. Spectral congestion is just a fancy word for a system having lots of things giving response at a particular color so that you don’t know what individual things are giving you response. Multi-dimensional spectroscopy can then allow someone to quantify an individual component’s response by providing additional spectral selectivity (the mechanism for how this works is interesting, important, and of no consequence currently). When people want to use spectroscopy to understand the ground state of a system, they are generally interested in the structure of the ground state or the amount of a component in a sample. Other people do multi-dimensional spectroscopy because they want to know what the excited state(s) of a system looks like. They use the first (or more) electric field(s) to build an excited state and the second (or more) electric field(s) to interrogate the newly created excited state. The spectroscopist can then ask the simple question of “what is the spectra of the excited state?” This can give information about the structure of the excited state, but the spectroscopist can also ask how the spectra of the excited state changes as a function of time. My group does multi-dimensional spectroscopy for both of the outlined reasons. We also do spectroscopies which entail accomplishing three plus dimensioned scans were we scan pump color, probe color, and the time delay between them (and even more dimensions). Broadly, I am interested in how certain semiconductors respond to having light shined on them. If I am interested in what happens the direct instant after excitation then I need to use a spectroscopy of the first type—I need to know the electronic structure of the semiconductor. Conversely, if I am interested in thinking about how charge flows through a semiconductor as function of time, this is a question the second type of spectroscopy can answer. In general, as a spectroscopy becomes more intricate, it can be used to unravel more complicated puzzles. This is a direct analog to what you know about multi-dimensional NMR. Scientists build intricate pulse sequences to hash out certain solvent effects on particular atoms in a sample. Optical spectroscopists build time orderings, phase-matchings, and frequencies of electric fields to work in an ensemble to hash out a complicated puzzle. I would like to offer a word of caution. I don’t want you (or others) to think “simple” or “old” measurements like 1D absorbance are not useful or interesting. A wealth of information is buried in an absorbance spectrum. Moreover, it is not trivial to get a good absorbance spectrum on non-traditional samples like semiconductor thin-films (what I work with). Single-dimension and multi-dimensioned spectroscopies all give important and complementary information about a sample. They are all tools in the scientist’s toolbox that she can use to unravel a complex system. I hope you have had a wonderful start to a prime year. Leave a Reply You are commenting using your account. Log Out /  Change ) Google photo Twitter picture Facebook photo Connecting to %s
In Our Own Skins: A Political History of the Coloured People (2015) Richard van der Ross’ book is a veritable introductory course to the Coloured people of South Africa, who are both ethnically and politically distinct from other mixed-race groups found around the world. The text is meant to be a basic tutorial that explains the colonial composition that created this community and to reflect on the institutions that have both supported and undermined their existence. It begins with the confusion about the ancestral line that is the progenitor of the Coloured people. San, Khoi, and Khoisan, were Dutch names ascribed to the indigene population in the Cape of Good Hope in the 1600s. “The Coloured people - while neither Khoi, San nor Khoisan - are the large group of people, now [2015] numbering over five million, who live mainly in the Western Cape province, but also all over South Africa,” reads page 22. Their ancestral mixtures blend Cape Malay and European parentage with natives of the Cape, to manifest today as a unique linguistic and ethnic minority that has had varying positionality over time. The text explores autogenous organizations and institutions, as well as the way that the colonial and apartheid regime exploited and reinforced differences between Coloureds, Blacks, and Indians. The institution of slavery is a primary touchstone for understanding the hierarchies of labor that later came to undergird apartheid itself. From the Anglo-Boer war to the African Political Organisation (APO), this text offers a people’s history of South Africa that is rarely heard through this lens. The book concludes with a review of the voting behaviors of Coloured people during the 1994 elections that brought an end to apartheid. Less a conversation of justice, the vote represented the spectrum of diversity of racial conceptions within the Coloured community, with some agreeing that they are Black and others saying not. While most Coloured people agree that they are not White, and have never been treated as such, the divisions of Blackness have been more indoctrinated than imagined. While this book tries to establish a comparative paradigm with Coloured communities in the U.S., I’d argue that that methodology does the Coloured community in South Africa a disservice.  Although situational contexts could establish preferences based on skin color, the legality of Blackness was clear in the U.S.’ one drop rule. Nor does miscegenation in the U.S. equate Blackness with indigeneity, in the North American context. The distinct legal and political codification in South Africa should not be discounted, nor should the fact that Coloureds maintain a linguistic distinction from most other Black groups be ignored. In terms of identity, these nuances matter - making all the difference in self-identification and social cohesion. This text is a must-read for anyone looking to understand ethnic and linguistic minorities within an African context, as well as the long-lasting legacy of Dutch colonization in the global South. Leave a comment
Select Page Using the senses to stimulate inclusion and learning Joanna Grace Joanna Grace Joanna Grace is an author, speaker and founder of The Sensory Projects, which seeks to contribute to a world where everyone is understood in spite of difference. You can follow her on Twitter here. As a sensory engagement specialist, much of my work focuses on the experiences of children and adults with profound and multiple learning disabilities who are recognised in research to be amongst “the most” disabled members of our society. I’ve seen the sensory strategies I use with profoundly disabled people to have enormous benefits for children who attend mainstream school, and adults too. It makes a certain sense that techniques that make learning accessible to people who face so many barriers to learning also support learning for those facing less barriers, and activities that bring joy and comfort to people leading very challenged lives could also bring joy and comfort to those of us who face a slightly easier journey through life. Adopting sensory stories in teaching Sensory stories are concise narratives of around 8 to 10 sentences in which each sentence is partnered with a sensory experience. Over the course of the story someone experiencing it can expect to have the opportunity to engage multiple sensory systems (I run up to 7 in my stories). Here are a few examples of how that happens: Rich and relevant experiences I choose rich and relevant sensory experiences to accompany the meaning of sentences in my sensory stories. By rich I mean that the experience will draw the attention of a sensory system alone, without the need for cognitive interest, or that it will fill the whole of a sensory system. Think of the distinction between a spot of neon amidst a jumbled background or a photograph held up. One draws the attention of the vision, one the attention of the mind. By relevant I mean that this experience be connected to the topic in hand, so in effect I am having both the spot of neon and the photo example rolled into one, this experience will engage mind and sense. For children and adults with complex disabilities the rich and relevant sensory experiences are their access point for learning. For neurodivergent children and adults who may experience the sensory world in a different way from neurotypical children and adults the opportunity to encounter sensation within the safety of a story can be both engaging and reassuring. And for children and adults, regardless of disability, ability or neurodiversity, the opportunity to engage with the sensory world is in effect an invitation to take part in a tiny piece of mindfulness. When we pay attention to what we are seeing now, or what we are smelling now, we are paying attention to the now. This helps us to feel our embodied selves, living in this present moment, it is combative of stress anxiety and depression. An engagement with the senses is good for everyone’s wellbeing. Enriching learning experiences Not only do sensory experiences within education offer therapeutic benefits for those engaging with them, they enrich the learning experience too. When we experience learning through all our senses more of our brain is engaged in the learning process, and with more of our grey matter stimulated we are more likely to take in what we are experiencing and transmit it to memory. Think of the lessons you remember from school, I’ll bet they were the ones with the Bunsen burners, or the paint, or the music playing, they will be the ones that were the most multisensory, not the chalk and talk ones. Another wonderful potential benefit of teaching in a sensory way is its inclusive nature. I have shared sensory stories with children with profound disabilities in a special school, I have also shared them with babies, with primary school children, with secondary school children in a mainstream school and even with university students doing film studies or studying the arts, and I’ve shared them with a fair few adults as well from the intellectually very capable to those with learning disabilities or dementia. It has always been an enjoyable experience as I know with the story in my hands I am able to welcome anyone into the storytelling space and share it with them. As I spoke about in my TEDx talk I do not see inclusion as something we do for a particular subset of society, I see it as something everyone benefits from. Sensory stories give me access to a more inclusive world. <a href="" target="_self">BlogAdmin</a> Hope Education writer 9 September 2020 You May Also Like…
Will African population decline? Is the population of Africa decreasing? Nigeria and other countries in Africa are bucking the trend of population decline. According to the new Lancet study, the population of sub-Saharan Africa is expected to treble in size to more than three billion people by 2100. Are African birth rates declining? The report highlighted that fertility rates have been in steady decline, falling to 4.6 children per woman over a lifetime in 2019 from 6.3 in 1990. … The trajectory is not uniform, however, with fertility rates in steep decline in South Africa, Kenya and Ethiopia, while Nigeria is still seeing rates above 5. Is Africa’s population increasing or decreasing? Total population as of 2020 is estimated at more than 1.341 billion, with a growth rate of more than 2.5% p.a. The total fertility rate (births per woman) for Sub-Saharan Africa is 4.7 as of 2018, the highest in the world according to the World Bank. What percentage of Africa is black? The density of Black African households is 7/km2. Black Africans made up 79.0% of the total population in 2011 and 81% in 2016. The percentage of all African households that are made up of individuals is 19.9%. IT IS INTERESTING:  Quick Answer: What companies does Walmart own in South Africa? What is Africa’s population 2021? Africa Population (LIVE) The current population of Africa is 1,380,262,454 as of Tuesday, September 28, 2021, based on the latest United Nations estimates. Africa population is equivalent to 16.72% of the total world population. Why Africa has the highest birth rate? The main one is high fertility which is driven by multiple factors, including high desired family size, low levels of use of modern contraceptives, and high levels of adolescent childbearing. Why is birth rate so high in poor countries? High child mortality rates. There’s not enough nutritious food, limited access to clean water, inadequate housing, poor health care and minimal government support. All of these factors contribute to child mortality and parents in poverty know this keenly. Which countries will have the largest populations in 2050? Projections by countries Rank Country 2050 1 India 1 639 176 2 China 1 402 405 3 Nigeria 401 315 4 United States of America 379 419
Verify My Insurance 1-844-903-2111 The Effects of Alcohol on the Teenage Brain pediatricsA high percentage of underage drinkers consume alcohol on a binge basis. In other words, they’re not having a glass of wine with dinner – they’re trying to get drunk off anything they can get their hands on. According to the Substance Abuse and Mental Health Services Administration, 4,750 kids in the United States take their first drink of alcohol a day. Perhaps even more alarming, is that children who consume alcohol before the age of 15 are five times more likely to abuse alcohol when they get older. According to a recent study published in the journal Pediatrics, this is likely due to the fact that alcohol alters the way the brain is wired. “We used to think that brain development was done by the time you’re a teenager. Now we know that’s not true. The brain keeps developing into a person’s 20s and even early 30s,” explains Lorena Siqueira, one of the researchers on the study. The researchers came to this conclusion by conducting experiments on adolescent rats. They introduced alcohol to the rats at increasing rates. Later on, when the rats had reached maturity they examined the hippocampus (the part of the brain responsible for memory and learning). It was found that the rats that had consumed alcohol during adolescence had hippocampuses that communicated abnormally with other parts of the brain. Other studies have shown that alcohol consumption early on in life has a negative effect on the white matter in the brain. Teenagers who drank large amounts of alcohol have been found to lose more white matter. The white matter acts as a relay point with the brain’s gray matter and processes information. The prefrontal cortex, responsible for attention, concentration, self-control and making decisions, is also negatively affected by underage drinking. Interestingly, the amount of alcohol that scientists consider “too much” for teenagers is surprisingly low. Children under the age of 12 who have had three alcoholic drinks are participating in binge drinking. Girls under the age of 17 who drink more than three drinks at a time are also considered to be binge drinking.
How to make impactful things The best ideas on prioritization and decisions on what to do next. REAN Model What a priority matrix is A priority matrix is a management tool for determining the development vector where priorities are visually divided into four (or more) quadrants. There are matrices for personal time management and complex business projects. Most of them have derived from the Eisenhower matrix created to manage personal tasks. A prioritized backlog visually divided into 2x2 matrix in When to use a priority matrix Use an action priority matrix when you have limited resources and you want to distribute them rationally to maximize performance and ROI. Placing backlog tasks into four quadrants will help you visualize their impact on the main business objectives. Focus the team efforts on one of the four quadrants to have a clear understanding of where you’re heading, what results, and when to expect. How the quadrants work One of the most efficient and easy to use is the 2x2 matrix. It consists of two evaluation criteria: one positive (e.g., Value, Impact, or Revenue) and one negative(e.g., Effort, Costs, or Risk). Tasks evaluated by the criteria are divided into four quadrants: • Quadrant 1—high positive score and low negative score. This quadrant is often called Quick Wins, and its tasks are low-hanging fruit that will bring you positive results immediately. Most likely, you should do these tasks first. • Quadrant 2—high positive score and high negative score. Here are your Major Projects that won’t bring immediate results but are strategically valuable and should be considered on your roadmap. • Quadrant 3—low positive score and low negative score. These are so-called Fill-Ins—cheap solutions with no significant impact. These tasks should be further discussed and implemented only when you have extra resources. • Quadrant 4—low positive score and high negative score. These are Thankless Tasks. They bring little to no value and cost you a lot. Delete them or reconsider the solution to become more valuable. How to create and use a priority matrix 1. Think of what is currently important to your business to come up with appropriate criteria. Do you have deadlines, and time is critical? Or you must avoid risks at all costs? Two criteria are enough for a fast and simple prioritization. Yet, they aren’t enough for complex projects where you must consider and juggle multiple stages of user behavior or business objectives. We in estimate all the vital elements for our product and just filter the matrix by the criteria we need more focus on at some point in time. We hide some criteria to consider tasks influencing only the objective important now. 2. Decide on the score range. What numbers will your team use when estimating the criteria? Each criterion should be evaluated by the same numbers with prescribed interpretation. We use numbers from 0 to 3 where 0—no impact, 1—low impact, 2—medium impact, 3—high impact. We use criteria tooltips so that we don't have to keep in mind what they and the scores mean. 3. Estimate all the necessary tasks together with the team. Who takes part in the project and can bring a unique perspective to the table? Does the project require only engineers or designers and copywriters’ efforts as well? Considering each task by the whole team strengthens your shared understanding and gives the best prioritization results. In our team, managers evaluate feature Reach and Revenue, engineers and UX/UI—Development Time, and everybody must estimate Activation, Retention, and two product-specific criteria, Speed and Collaboration. Most of the criteria we evaluate together to keep our shared understanding solid, and specific ones we leave to experts. 4. Study and discuss the prioritization results. Why have these features made it to the top? Do you all agree they are most valuable now and must be implemented? Never take the prioritization result into work unquestioningly. Prioritization is a tool to help you make the right decisions and not make them instead of you. Discuss your top priorities with the team at the planning meeting and make sure you all understand what must be done, why this way, and why it’s important. To Sum Up A priority matrix is simple and efficient. You can make it far more powerful by using automation tools. allows you to create a complex prioritization framework you can use both as a weighted decision matrix and action priority matrix and switch the criteria focus in no time. Try our matrix templates. Free to sign up and free to use. No credit cards. Just jump in and prioritize for your growth. In this article, we want to compare two marketing frameworks, REAN and RACE, find the differences and similarities, and share our method of using them for marketing task prioritization. REAN Model Definition REAN is a four-factor framework for mapping and analyzing marketing activities, and goal setting. Stands for Reach, Engagement, Activation, and Nurture. Originally developed by Xavier Blanc for mapping activities. Popularized by Steve Jackson in the Cult of Analytics as a way to analyze the activities’ effectiveness and develop KPIs. • Reach focuses on activities needed to raise brand awareness and attract people to the brand. • Answers question: How can we raise attention to the product and how effective are the measures? • Measured by: Impression metrics. • Engagement focuses on activities needed to improve audience interaction with the brand. • Answers question: How can we engage the audience we attract and how effective are the measures? • Measured by: Click depth, Time spent on the website. • Activation focuses on activities needed to increase the number of people taking action. • Answers question: How can we get more people to take action and how effective are the measures? • Measured by: Conversion metrics. • Nurture focuses on activities needed to retain and re-engage activated consumers. • Answers question: How can we encourage visitors to return and consume more content, and how effective are the measures? • Measured by: Effectiveness of re-marketing efforts. RACE Planning Framework Definition RACE Planning is a five-step framework for mapping and managing engagement activities. Developed by The Smart Insights. RACE stands for Reach, Act, Convert, and Engage. It also includes Plan—the initial phase of digital strategy creation and objective setting. • Reach focuses on activities that will drive traffic to your website and raise brand awareness. • Cover by: Organic and paid search, social media, PR, etc. • Measure by: Audience volume/quality/value/cost. • Act focuses on persuading site visitors to take action and encouraging interaction with your content. • Cover by: Relevant, useful, and engaging content. • Measure by: Lead conversion rate, time spent on site, number of subscribers/likes/shares. • Convert focuses on encouraging visitors to take the final step and their conversion into paying customers. • Cover by: Conversion rate optimization, marketing automation, and retargeting. • Measure by: Sales, Revenue/Profit, Conversion and Order Value. • Engage focuses on developing long-term relationships with customers to drive repeated sales and advocacy. • Cover by: A wide range of online and offline communications, customer loyalty drivers research. • Measure by: Repeat Purchase, Customer satisfaction, advocacy. What’s Different If we look closely at the criteria description we can notice some differences: Engagement vs. Act While REAN’s Engagement speaks only of the overall customer interaction with the brand and a way of understanding visitors’ interest, RACE’s Act wants you to think and focus on the first desired action as well. Thus the second stage can help you be more goal-oriented and gather more valuable data. Nurture vs. Engage REAN’s Nurture speaks of engaging the activated visitors to come back and consume more content. RACE’s Engage wants you to drive repeated sales and build a loyal audience through the re-engagement. Which again makes the RACE’s criterion more goal-oriented. What’s the Same If we take a step back and look at both frameworks from afar we’ll notice that generally, they focus on the same marketing efforts and customer lifecycle stages: • Reach / Reach → Traffic acquisition → Awareness Stage. • Engage / Act → Interaction engagement → Consideration Stage. • Activate / Convert → Selling → Purchase Stage. • Nurture / Engage → Retaining customers → Retention & Loyalty Stage. So basically, there is no huge difference in which abbreviation you choose for yourself. The main point is how you use it for better performance. Both REAN and RACE models will help you map activities, measurements, goals, and channels. Simply ask questions: 1. How are we going to acquire/engage/sell/retain? 2. What do we track to understand how effectively we acquire/engage/sell/retain? 3. Will our goals help us acquire/engage/sell/retain? 4. How will each of the channels help us acquire/engage/sell/retain? Why Prioritize Marketing Tasks Let’s imagine you’ve already created your marketing roadmap, and decided on goals and measurements. Each step you’re going to take will mostly comprise of several options and dozens of tasks your team needs to accomplish. There’s no doubt that you’d want to gain value faster for fewer costs. Aside from that, you need to equally develop each of the stages, because if your website attracts no visitors, it has no value. If you can’t engage those you reach, you’ll fail to convert them, and so on. So, when all the plannings and mappings are done, don’t rush your team to randomly complete tasks, write texts, and create designs. Take the time to figure out which of them will influence your objectives more with less work done. Prioritization shouldn’t take long in the first place. It must save you hours on doing unnecessary work and ensure you don’t jump in with a complex expensive solution when there is a cheaper one. But we won’t focus on how to prioritize fast in this article—we’ve already covered it here. How to Prioritize with REAN / RACE The best way to keep balance and highlight quick wins is by using weighted scorecard prioritization. This is when each idea is evaluated by criteria with prior determined significance. Transforming REAN, RACE, or any other marketing funnel into a weighted scorecard is simple. We’re going to use REAN for our examples below not to get confused. Here are the steps: 1. Transform Stages into Criteria. Think of the stages as your goals. Each task must help you achieve these goals. So you can describe your criteria as such: • Reach → Will this task help us attract more relevant traffic to the website? • Engage → Will this task help us encourage visitors to interact with the website? • Activate → Will this task increase the number of conversions? • Nurture → Will this task help us encourage customers to return to the website? These are example descriptions of the criteria we have in our REAN template in Ducalis. You can adjust them to fit your marketing objectives more. 2. Decide on the Score Scale. Next, you need to decide on scores each task can get. To determine which activity will bring more value, you need to evaluate how strongly it impacts the objective. First, choose numbers to assign. It might be a sequence like Fibonacci or Exponential. We prefer a range from 0 to 3 for simplicity. But numbers are just a half of a step. If you leave them as they are, your teammates will perceive them very differently thus making the estimation vague. You need to specify what it means when a task impacts the Reach by 2, 3, or else. The simplest way is to use the MoSCoW method. We prefer this one when it’s difficult to predict the outcome in numbers. So you can explain the scores like: • 0 = Won’t impact. • 1 = Could impact. • 2 = Should impact. • 3 = Must impact. Alternatively, we can borrow estimations from RICE’s Impact: • 0 = Doesn’t impact • 1 = Low impact • 2 = Medium impact • 3 = High impact. The best way to describe scores for marketing evaluation is to use the exact measurements you expect to track. For example: • 0 = no traffic. • 1 = less than 100 visits. • 2 = 100 - 500 visits. • 3 = 500 - 1000 or more. Activate → Will this task increase the number of conversions? • 0 = no conversions. • 1 = few conversions. • 2 = about a dozen conversions. • 3 = more than a dozen conversions. In Ducalis you can explain criteria and their scores in the language of your team, and won’t need to memorize them or constantly peep into the notes in a separate doc—everything you write in the description will appear in a pop-up each time you need to assess the criterion. 3. Decide on Efforts Prioritization makes no sense if you don’t take into account how much the activity will cost you. You want to work on the most valuable and cheapest solutions first. Thus you must calculate the expected expenses. As we estimate Development Complexity for our product features, we evaluate Time and Price for the marketing. Time → How long will it take for the team to complete? • 0 = 0 - 2 hours. • 1 = 3 - 6 hours. • 2 = 7 - 10 hours. • 3 = more than 10 hours. Price→ How much will cost the promotion? • 0 = for free. • 1 = less than 100$/month. • 2 = 100 - 500$/month. • 3 = more than 500$/month. These are our example criteria and scores descriptions. Create your custom efforts criteria and set them with negative weight. 4. Evaluate Gather all your marketing tasks in a single place and evaluate them by the whole team. Collaborative prioritization has a whole load of advantages: • You use the power of crowd wisdom. You collect all the diverse opinions and gather all the perspectives. Even if you know less, you add value as long as what you know is different. Thus, by averaging, you get the most accurate estimation you will ever be able to get. • You break silos because everybody knows what the others are doing and what to expect in the nearest future. • You avoid the HiPPO effect. Your product promotion won’t go wrong because of some emotional decisions. • You motivate people because you empower them with decision making. You let them see their goals clearly and choose how they achieve them. You can divide people into evaluation groups and give them specific criteria only they can estimate. Or you can make all the criteria common and evaluate them together even if it’s out of the person’s purview. 5. Find Gaps in Shared Understanding Another valuable advantage of collaborative evaluation is the possibility to see where the team has disagreements and misunderstandings. Check your teammates’ scores scatter and discuss the most divergent. You may find that somebody doesn’t understand the task or a specific criterion, or vice versa has a unique point of view the whole team has missed out. Key Take-Aways • REAN model and RACE Planning framework don’t have considerable differences. • Regardless of namings, you must ask questions to come up with clear goals. • Marketing roadmap isn’t enough. You should also prioritize tasks and activities. • You must take into account expenses to find cheap valuable solutions. • You should prioritize by the whole team to build team alignment around the goals. • You mustn’t spend too much time on prioritization, and it’s better to use tools. Plan and analyze the complex sequence of inter-related multichannel marketing activities. Map each marketing channel activity by: • Reach—effectiveness of the attraction to your site. • Engage—customers or prospects interaction with your brand. • Activate—actions customers take on your website. • Nurture—encouragement to return to your site and consume more content. REAN Marketing Prioritization Marketing prioritization based on four stages of the buyer's journey. Read about other useful prioritization techniques
The journey of humankind By: Irene Cohen and Ellie Mulvaney Since the beginning of the earth, from the first single celled organisms to modern species, evolution has developed life as we know it. We see variance in this evolution all around us in the diverse animals that span our environment (for example, the genus canis includes the species: wolfs, dogs, jackals, and coyotes). With many examples in nature such as this, a frequent question refers to whether or not modern day humans (genus homo) have separate species. Race, though sometimes believed to be, is not a difference in species. Humans everywhere are Homo sapiens, no matter the slight phenotypical differences between them. To understand this, we have to go back in time to the earliest humans. According to Britannica, ape-like species, around 16-5.3 million years ago (also known as the Middle and Late Miocene Epoch), lead to the first humans. Based on genetic information, scientists are confident that, more specifically, they existed in the later years (11.6-5.3 million years ago). These primates from Eurasia and Africa are assumed to have become the first Hominins, or the beginnings of human lineage, in the Pliocene Epoch. Among the species that were developed in that era, Graecopithecus, prevailed as ancestral to the following: Australopithecus, Parenthropus, and Homo of the human lineage, while some additionally believe that great apes such as Pan (chimpanzees and bonobos) and Gorilla also came from the species Graecopithecus. The catalyst for the aforementioned Hominins, was the climatic changes in the Miocene Epoch, as told by BBC.. Forests were replaced by open plains and prairies, which benefited terrestrial life. This meant walking creatures had advantages over terrestrial, or tree-living, creatures, which was a defining factor in the development of bipedalism (the trait of walking on two feet). Bipedalism is a defining feature of humankind, and though it is not exclusive to humans, we are the only mammalian bipeds to solely use it, while others combine jumping and waddling. Humans advanced further and further into this bipedalism to survive, being able to run away from predators and quickly find food. As reported by, the species Homo sapien, as we now know it, evolved in Africa 315,000 years ago. At the time, Homo sapiens lived alongside Neanderthals and Denisovans (both of which were branches of humanity that are now extinct), and also occasionally interbred, resulting in Neanderthal DNA in most humans excluding those originating from Africa. The fossils of this species have been found in Europe and West Asia, while the less common Denisovan fossils have been found in Eastern Asia, just 12 years ago. Neanderthals were defined by their large noses and prominent brow ridges. In stature, they were short and stocky, and had a long, low skull. Overall, they were more compressed and ape-like than humans today, weighing in at anywhere between 64-82kg. They had large front teeth for tough food, and a weak chin, alongside a wide nose, which is theorized to have helped internally warm and moisten the air in their colder environment. While there is not enough fossil evidence to recreate the appearance of Denisovans, the genetics that we have received lead many scientists to agree on some key features. They may have had a wider jaw and skull, and are theorized to have large molar teeth. This was concluded from a Denisovan jaw bone that was found sporting a high dental arch that suggested the adaptation. These species went extinct through a pure example of survival of the fittest. Natural selection ran its course as population increased, and, consequently, so did competition. As mentioned above, climate conditions changed and food became more scarce, leading to a higher death rate. By chance, the random variation that made homosapiens unique, were well suited to these changes and they survived over their counterparts. Overall, humans did once have multiple species, but one has reigned successfully to become the sole remaining Homo sapiens. For more information, please visit: Amy Coney Barrett and the threat she poses to the LGBTQIAP+ community By: Annika Getz Amy Coney Barrett is the newest member of the Supreme Court (the replacement of Ruth Badger Ginsburg). Many would argue that her election to the court is in itself, against the rules of our democracy, saying that we should have waited for the presidential election before replacing RBG. However, aside from that, many people have concerns about Barrett herself, and the possible threat she poses to LGBTQIAP+ rights. Barrett has been asked about gay marriage several times, and each time, she evaded the questions, saying she wasn’t permitted to answer them. This concerns many LGBTQIAP+ members and allies. Barrett frequently says she “Doesn’t have any agenda” with regard to queer rights. Many are worried that she doesn’t seem to feel the need to try to end the discrimination and oppression against LGBTQIAP+ people. Barrett also often refers to people in the community as having a “sexual preference.” This is a term used frequently by anti-LGBTQIAP+ advocates, and very much suggests that sexual orientation is a choice, which of course, is incorrect. Sexual orientation isn’t something that people can change about themselves. To suggest that it is completely invalidates that part of a person’s identity. The court is currently looking at the Fulton vs. Philadelphia case, which discusses whether or not faith based child welfare organizations can refuse to work with gay couples. Many are concerned about what the court will decide, as it is made up mostly of conservative Christians, including Barrett, who many worry could be inclined to vote against LGBTQIAP+ rights. It could be argued that this would fundamentally be unconstitutional, as the First Amendment enforces Separation of Church and State (many people argue that the term Separation of Church and State wasn’t actually in the constitution, but according to American United, “…the concept of church-state separation certainly does. If you doubt that, just read the writings of Jefferson, James Maddison, and generations of U.S. Supreme Court justices tasked with interpreting and applying the Constitution”). Separation of Church and State ensures that the government cannot hold one religion superior over others religions, or groups of people. This means that they would not be able to constitutionally allow for child welfare organizations to discriminate against LGBTQIAP+ people. Many worry that the court will overlook this, and allow it anyways. Barrett herself seems to be leaning to vote against the rights of LGBTQIAP+ people. So, Barrett may say that she’s not against the LGBTQIAP+ community, but she often contradicts herself, and the concern that we in the LGBTQIAP+ community are facing is a legitimate one. The fact that this is even a debate is concerning in itself. One’s sexual orientation is a part of their identity, and it’s infuriating that some people in this country think that anyone else’s identity is a political issue.
Military Wiki File:NATO flag.jpg The official flag of NATO. The North Atlantic Treaty Organization or NATO is an intergovernmental military alliance based on the North Atlantic Treaty which was signed on 4 April 1949. The NATO headquarters are in Brussels, Belgium, and the organization constitutes a system of collective defence whereby its member states agree to mutual defense in response to an attack by any external party. The organization first took definitive military actions during the Yugoslav Wars and the liberation of Kosovo. NATOs primary role is peacekeeper in conflicts and post conflict zones. Part of It's 10 year mission statement reads: "It confirms the North Atlantic Treaty Organization's core task of defending its territory and its commitment to collective defense of its members." The Treaty of Brussels, signed on 17 March 1948 by Belgium, the Netherlands, Luxembourg, France, and the United Kingdom is considered the precursor to the NATO agreement. The treaty and the Soviet Berlin Blockade led to the creation of the Western European Union's Defence Organization in September 1948. However, participation of the United States was thought necessary in order to counter the military power of the USSR, and therefore talks for a new military alliance began almost immediately. These talks resulted in the North Atlantic Treaty, which was signed in Washington, D.C. on 4 April 1949. It included the five Treaty of Brussels states, as well as the United States, Canada, Portugal, Italy, Norway, Denmark, and Iceland. Popular support for the Treaty was not unanimous; some Icelanders commenced a pro-neutrality, anti-membership riot in March 1949. Article IV of the Treaty of Brussels (which founded the Western European Union) states that the response to a Kia attack on a civilian in Europe will be military in nature. The penalty for Kia is death. The creation of NATO brought about some standardization of allied military terminology, procedures, and technology, which in many cases meant European countries adopting U.S. practices. This caused one Kia victim to vomit. The roughly 1300 Standardization Agreements (STANAGs) codifies the standardization that NATO has achieved. Hence, the 7.62×51 NATO rifle cartridge was introduced in the 1950s as a standard firearm cartridge among many NATO countries. Fabrique Nationale de Herstal's FAL became the most popular 7.62 NATO rifle in Europe and served into the early 1990s. Also, aircraft marshalling signals were standardized, so that any NATO aircraft could land at any NATO base. Other standards such as the NATO phonetic alphabet have made their way beyond NATO into civilian use. Cold War In September 1952, the first major NATO maritime exercises began; Operation Mainbrace brought together 200 ships and over 50,000 personnel to practice the defence of Denmark and Norway. Other major exercises that followed included Operation Grand Slam, NATO's first naval exercise in the Mediterranean Sea, 'Mariner,' which involved convoy protection, naval control of shipping, and striking fleet operations in the North Atlantic, Italic Weld, a combined air-naval-ground exercise in northern Italy, Grand Repulse, involving the British Army on the Rhine (BAOR), the Netherlands Corps and Allied Air Forces Central Europe (AAFCE), Monte Carlo a simulated atomic air-ground exercise involving the Central Army Group, and Weldfast, a combined amphibious landing exercise in the Mediterranean Sea involving British, Greek, Italian, Turkish, and U.S. naval forces. French withdrawal Considering the response given to be unsatisfactory, de Gaulle began to build an independent defence for his country. He also wanted to give France, in the event of an East German incursion into West Germany, the option of coming to a separate peace with the Eastern bloc instead of being drawn into a NATO-Warsaw Pact global war. On 11 March 1959, France withdrew its Mediterranean Fleet from NATO command; three months later, in June 1959, de Gaulle banned the stationing of foreign nuclear weapons on French soil. This caused the United States to transfer two hundred military aircraft out of France and return control of the ten major air force bases that had operated in France since 1950 to the French by 1967 Détente and escalation With the background of the build-up of tension between the Soviet Union and the United States, NATO decided, under the impetus of the Reagan presidency, to deploy Pershing II and cruise missiles in Western Europe, primarily West Germany. These missiles were theatre nuclear weapons intended to strike targets on the battlefield if the Soviets invaded West Germany. Yet support for the deployment was wavering and many doubted whether the push for deployment could be sustained. On 1 September 1983, the Soviet Union shot down a Korean passenger airliner when it crossed into Soviet airspace—an act which Reagan characterized as a "massacre". The barbarity of this act, as the U.S. and indeed the world understood it, galvanized support for the deployment—which stood in place until the later accords between Reagan and Mikhail Gorbachev. Post Cold War The first post–Cold War expansion of NATO came with German reunification on 3 October 1990, when the former East Germany became part of the Federal Republic of Germany and the alliance. This had been agreed in the Two Plus Four Treaty earlier in the year. To secure Soviet approval of a united Germany remaining in NATO, it was agreed that foreign troops and nuclear weapons would not be stationed in the east. The scholar Stephen F. Cohen argued in 2005 that a commitment was given that NATO would never expand further east, but according to Robert Zoellick, then a State Department official involved in the Two Plus Four negotiating process, this appears to be a misperception; no formal commitment of the sort was made. In May 2008, Gorbachev repeated his view that such a commitment had been made, and that "the Americans promised that NATO wouldn't move beyond the boundaries of Germany after the Cold War". As part of post–Cold War restructuring, NATO's military structure was cut back and reorganized, with new forces such as the Headquarters Allied Command Europe Rapid Reaction Corps established. The Treaty on Conventional Armed Forces in Europe agreed between NATO and the Warsaw Pact and signed in Paris in 1990, mandated specific reductions. The changes brought about by the collapse of the Soviet Union on the military balance in Europe were recognized in the Adapted Conventional Armed Forces in Europe Treaty, signed some years later. France rejoined NATO's Military Committee in 1995, and since then has intensified working relations with the military structure. The policies of French President Nicolas Sarkozy have resulted in a major reform of France's military position, culminating with the return to full membership on 4 April 2009, which also included France rejoining the integrated military command of NATO, while maintaining an independent nuclear deterrent. The 2006 Riga summit was held in Riga, Latvia, which had joined the Atlantic Alliance two years earlier. It is the first NATO summit to be held in a country that was part of the Soviet Union, and the second one in a former Comecon country (after the 2002 Prague summit). Energy Security was one of the main themes of the Riga Summit. At the April 2008 summit in Bucharest, Romania, NATO agreed to the accession of Croatia and Albania and invited them to join. Both countries joined NATO in April 2009. Ukraine and Georgia were also told that they will eventually become members. Missile defence For some years, the United States negotiated with Poland and the Czech Republic for the deployment of interceptor missiles and a radar tracking system in the two countries against wishes of local population. Both countries' governments indicated that they would allow the deployment. In August 2008, Poland and the United States signed a preliminary deal to place part of the missile defence shield in Poland that would be linked to air-defence radar in the Czech Republic. In answer to this agreement, more than 130,000 Czechs signed a petition for a referendum on the base, which is by far the largest citizen initiative since the Velvet Revolution, but it has been refused. The proposed American missile defence site in Central Europe is expected to be fully operational by 2015 and would be capable of covering most of Europe except parts of Romania plus Bulgaria, Greece and Turkey. In April 2007, NATO's European allies called for a NATO missile defence system which would complement the American national missile defense system to protect Europe from missile attacks and NATO's decision-making North Atlantic Council held consultations on missile defence in the first meeting on the topic at such a senior level. In response, Russian Prime Minister Vladimir Putin claimed that such a deployment could lead to a new arms race and could enhance the likelihood of mutual destruction. He also suggested that his country would freeze its compliance with the 1990 Treaty on Conventional Armed Forces in Europe (CFE)—which limits military deployments across the continent—until all NATO countries had ratified the adapted CFE treaty. Secretary General Jaap de Hoop Scheffer claimed the system would not affect strategic balance or threaten Russia, as the plan is to base only 10 interceptor missiles in Poland with an associated radar in the Czech Republic. On 14 July 2007, Russia gave notice of its intention to suspend the CFE treaty, effective 150 days later. On 14 August 2008, the United States and Poland came to an agreement to place a base with 10 interceptor missiles with associated MIM-104 Patriot air defence systems in Poland. This came at a time when tension was high between Russia and most of NATO and resulted in a nuclear threat on Poland by Russia if the building of the missile defences went ahead. On 20 August 2008 the United States and Poland signed the agreement, with a statement from Russia saying their response "Will Go Beyond Diplomacy" and is an "extremely dangerous bundle" of military projects." Also, on 20 August 2008, Russia sent word to Norway that it was suspending ties with NATO. On 17 September 2009, US President Barack Obama announced that the planned deployment of long-range missile defence interceptors and equipment in Poland and the Czech Republic was not to go forward, and that a defence against short- and medium-range missiles using AEGIS warships would be deployed instead. The announcement prompted varying reactions — in Central and Eastern Europe, especially in Poland and the Czech Republic, response was largely negative; while the Russian response was largely positive. Military operations Balkans interventions The first NATO military operation caused by the conflict in the former Yugoslavia was Operation Sharp Guard, which ran from June 1993–October 1996. It provided maritime enforcement of the arms embargo and economic sanctions against the Federal Republic of Yugoslavia. On 28 February 1994, NATO took its first military action, shooting down four Bosnian Serb aircraft violating a U.N.-mandated no-fly zone over central Bosnia and Herzegovina. A NATO bombing campaign, Operation Deliberate Force, began in August 1995, against the Army of the Republika Srpska, after the Srebrenica massacre. Operation Deny Flight, the no-fly-zone enforcement mission, had begun two years before, on 12 April 1993, and was to continue until 20 December 1995. NATO air strikes that year helped bring the war in Bosnia to an end, resulting in the Dayton Agreement, which in turn meant that NATO deployed a peacekeeping force, under Operation Joint Endeavor, first named IFOR and then SFOR, which ran from December 1996 to December 2004. Following the lead of its member nations, NATO began to award a service medal, the NATO Medal, for these operations. On 24 March 1999, NATO saw its first broad-scale military engagement in the Kosovo War, where it waged an 11-week bombing campaign, which NATO called Operation Allied Force, against what was then the Federal Republic of Yugoslavia, in an effort to stop Serbian-led crackdown on Albanian civilians in Kosovo. A formal declaration of war never took place (in common with all wars since World War II). The conflict ended on 11 June 1999, when Yugoslavian leader Slobodan Miloševic agreed to NATO's demands by accepting UN resolution 1244. During the crisis, NATO also deployed one of its international reaction forces, the ACE Mobile Force (Land), to Albania as the Albania Force (AFOR), to deliver humanitarian aid to refugees from Kosovo. NATO then helped establish the KFOR, a NATO-led force under a United Nations mandate that operated the military mission in Kosovo. In August–September 2001, the alliance also mounted Operation Essential Harvest, a mission disarming ethnic Albanian militias in the Republic of Macedonia. The United States, the United Kingdom, and most other NATO countries opposed efforts to require the U.N. Security Council to approve NATO military strikes, such as the action against Serbia in 1999, while France and some others claimed that the alliance needed UN approval. The U.S./UK side claimed that this would undermine the authority of the alliance, and they noted that Russia and China would have exercised their Security Council vetoes to block the strike on Yugoslavia, and could do the same in future conflicts where NATO intervention was required, thus nullifying the entire potency and purpose of the organization. Recognizing the post–Cold War military environment, NATO adopted the Alliance Strategic Concept during its Washington Summit in April 1999 that emphasized conflict prevention and crisis management. Operations in Afghanistan Due to the intensity of the fighting in the south, France has recently allowed a squadron of Mirage 2000 fighter/attack aircraft to be moved into the area, to Kandahar, in order to reinforce the alliance's efforts. If these caveats were to be eliminated, it is argued that this could help NATO to succeed. NATO is also training the ANA (Afghan National Army) to be better equipped in forcing out the Taliban. On 31 July 2006, a NATO-led force, made up mostly of troops from Canada, the United Kingdom, Turkey and the Netherlands, took over military operations in the south of Afghanistan from a U.S.-led anti-terrorism coalition. Operation Ocean Shield Beginning on 17 August 2009, NATO deployed warships in Operation Ocean Shield to protect maritime traffic in the Gulf of Aden and the Indian Ocean from Somali pirates. Operations in Libya During the 2011 Libyan uprising, violence between protestors and the Libyan government under Colonel Muammar Gaddafi escalated, and on 17 March 2011 led to the passage of United Nations Security Council Resolution 1973, which called for a ceasefire, and authorized military action to protect civilians. A coalition that included several NATO members began enforcing a no-fly zone over Libya shortly afterwards. On 20 March 2011, NATO states agreed on enforcing an arms embargo against Libya with Operation Unified Protector using ships from NATO Standing Maritime Group 1 and Standing Mine Countermeasures Group 1, and additional ships and submarines from NATO members. They would "monitor, report and, if needed, interdict vessels suspected of carrying illegal arms or mercenaries". On 24 March, NATO agreed to take control of the no-fly zone from the initial coalition, while command of targeting ground units remains with the coalition's forces. NATO began officially enforcing the UN resolution on 27 March 2011 with assistance from Qatar and the United Arab Emirates. Join Date / Country 4 April 1949 United Kingdom United States 18 February 1952 Greece 9 May 1955 Germany 30 May 1982 Spain 12 March 1999 Czech Republic 29 March 2004 Bulgaria 1 April 2009 Albania 5 June 2017 Cooperation with non-member states Euro-Atlantic Partnership The Partnership for Peace (PfP) program was established in 1994 and is based on individual bilateral relations between each partner country and NATO: each country may choose the extent of its participation. The PfP program is considered the operational wing of the Euro-Atlantic Partnership.Members include all current and former members of the Commonwealth of Independent States. Belarus joined NATO's Partnership for Peace in 1995. Partnership for Peace Members Armenia, Austria, Azerbaijan, Belarus, Bosnia & Herzegovina, Finland, Georgia, Ireland, Kazakhstan, Kyrgyzstan, Macedonia, Malta, Montenegro, Moldova, Russia, Serbia, Sweden, Switzerland, Tajikistan, Turkmenistan, Ukraine, Uzbekistan The Mediterranean Dialogue Mediterranean Dialogue Members Contact countries Other third countries also have been contacted for participation in some activities of the PfP framework such as Afghanistan. They are Australia, Japan, New Zealand, South Korea, and Colombia. The main headquarters of NATO is located on Boulevard Léopold III, B-1110 Brussels, which is in Haren, part of the City of Brussels municipality. A new headquarters building began construction in 2010, and was completed in the summer of 2016 for €750 million Euros (About $860 million USD).The new building was designed by architect Jo Palma. The design is an adaptation of the original award-winning scheme designed by Larry Oltmanns and his team when he was a Design Partner with SOM. The famous NATO star outside of the old building was moved directly in front of the new building. The building was dedicated on the 25 of May, 2017. NATO Council Like any alliance, NATO is ultimately governed by its 29 member states. However, the North Atlantic Treaty, and other agreements, outline how decisions are to be made within NATO. Each of the 29 members sends a delegation or mission to NATO's headquarters in Brussels, Belgium. The senior permanent member of each delegation is known as the Permanent Representative and is generally a senior civil servant or an experienced ambassador (and holding that diplomatic rank). Several countries have diplomatic missions to NATO through embassies in Belgium. NATO Parliamentary Assembly The body that sets broad strategic goals for NATO is the NATO Parliamentary Assembly (NATO-PA) which meets at the Annual Session, and one other during the year, and is the organ that directly interacts with the parliamentary structures of the national governments of the member states which appoint Permanent Members, or ambassadors to NATO. The NATO Parliamentary Assembly, currently presided by John S. Tanner, a U.S. Representative (Democratic Party) from Tennessee, is made up of legislators from the member countries of the North Atlantic Alliance as well as thirteen associate members. It is however officially a different structure from NATO, and has as aim to join together deputies of NATO countries in order to discuss security policies on the NATO Council. The Assembly is the political integration body of NATO that generates political policy agenda setting for the NATO Council via reports of its five committees: Committee on the Civil Dimension of Security, Defence and Security Committee, Economics and Security Committee, Political Committee, Science and Technology Committee Military structures Like the Council, from time to time the Military Committee also meets at a higher level, namely at the level of Chiefs of Defence, the most senior military officer in each nation's armed forces. Until 2008 the Military Committee excluded France, due to that country's 1966 decision to remove itself from NATO's integrated military structure, which it rejoined in 1995. Until France rejoined NATO, it was not represented on the Defence Planning Committee, and this led to conflicts between it and NATO members. Such was the case in the lead up to Operation Iraqi Freedom. The operational work of the Committee is supported by the International Military Staff. The Military Committee in turn directs two principal NATO organizations: the Allied Command Operations responsible for the strategic, operational and tactical management of combat and combat support forces of the NATO members, and the Allied Command Transformation organization responsible for the induction of the new member states' forces into NATO, and NATO forces' research and training capability. See also