4. Quantitative Analysis of Leadership Decapitation in Terrorist Groups

2019 ◽  
pp. 103-149
2012 ◽  
Vol 36 (4) ◽  
pp. 9-46 ◽  
Author(s):  
Bryan C. Price

Several states, including Israel and the United States, have put decapitation tactics, which seek to kill or capture leaders of terrorist organizations, at the forefront of their counterterrorism efforts. The vast majority of scholarly work on decapitation suggests, however, that leadership decapitation is ineffective at best and counterproductive at worst. Contrary to this conventional wisdom, leadership decapitation significantly increases the mortality rate of terrorist groups, although the results indicate that the effect of decapitation decreases with the age of the group, even to a point where it may have no effect at all. This finding helps to explain the previously perplexing mixed record of decapitation effectiveness. Terrorist groups are especially susceptible to leadership decapitation because their organizational characteristics (they are violent, clandestine, and values based) amplify the difficulties of leadership succession. Additionally, in contrast to the conventional wisdom regarding the durability of terrorist groups, politically relevant terrorist groups (defined as those with at least four attacks including one attack resulting in a fatality) endure significantly longer than previously believed.


Author(s):  
Jenna Jordan

Leadership decapitation, which refers to the arrest or killing of a group’s leadership, has become a key tool in current counterterrorism strategies This chapter presents an overview of the debate about leadership decapitation against terrorist organizations. Much of the optimism regarding the efficacy of decapitation as a counterterrorism strategy stems from the belief that terrorist groups depend upon the charisma of leaders for their cohesion. In order to understand whether decapitation works, this chapter presents three questions that will be answered in the book: (1) Under what conditions does leadership decapitation result in the decline of a terrorist organization? (2) Does leadership decapitation lengthen or shorten a group’s life span? (3) In cases where decapitation does not result in a group’s collapse, to what extent does it weaken a group and hinder its capacity to carry out attacks?


Author(s):  
J.P. Fallon ◽  
P.J. Gregory ◽  
C.J. Taylor

Quantitative image analysis systems have been used for several years in research and quality control applications in various fields including metallurgy and medicine. The technique has been applied as an extension of subjective microscopy to problems requiring quantitative results and which are amenable to automatic methods of interpretation.Feature extraction. In the most general sense, a feature can be defined as a portion of the image which differs in some consistent way from the background. A feature may be characterized by the density difference between itself and the background, by an edge gradient, or by the spatial frequency content (texture) within its boundaries. The task of feature extraction includes recognition of features and encoding of the associated information for quantitative analysis.Quantitative Analysis. Quantitative analysis is the determination of one or more physical measurements of each feature. These measurements may be straightforward ones such as area, length, or perimeter, or more complex stereological measurements such as convex perimeter or Feret's diameter.


Author(s):  
V. V. Damiano ◽  
R. P. Daniele ◽  
H. T. Tucker ◽  
J. H. Dauber

An important example of intracellular particles is encountered in silicosis where alveolar macrophages ingest inspired silica particles. The quantitation of the silica uptake by these cells may be a potentially useful method for monitoring silica exposure. Accurate quantitative analysis of ingested silica by phagocytic cells is difficult because the particles are frequently small, irregularly shaped and cannot be visualized within the cells. Semiquantitative methods which make use of particles of known size, shape and composition as calibration standards may be the most direct and simplest approach to undertake. The present paper describes an empirical method in which glass microspheres were used as a model to show how the ratio of the silicon Kα peak X-ray intensity from the microspheres to that of a bulk sample of the same composition correlated to the mass of the microsphere contained within the cell. Irregular shaped silica particles were also analyzed and a calibration curve was generated from these data.


Author(s):  
H.J. Dudek

The chemical inhomogenities in modern materials such as fibers, phases and inclusions, often have diameters in the region of one micrometer. Using electron microbeam analysis for the determination of the element concentrations one has to know the smallest possible diameter of such regions for a given accuracy of the quantitative analysis.In th is paper the correction procedure for the quantitative electron microbeam analysis is extended to a spacial problem to determine the smallest possible measurements of a cylindrical particle P of high D (depth resolution) and diameter L (lateral resolution) embeded in a matrix M and which has to be analysed quantitative with the accuracy q. The mathematical accounts lead to the following form of the characteristic x-ray intens ity of the element i of a particle P embeded in the matrix M in relation to the intensity of a standard S


Author(s):  
John A. Hunt

Spectrum-imaging is a useful technique for comparing different processing methods on very large data sets which are identical for each method. This paper is concerned with comparing methods of electron energy-loss spectroscopy (EELS) quantitative analysis on the Al-Li system. The spectrum-image analyzed here was obtained from an Al-10at%Li foil aged to produce δ' precipitates that can span the foil thickness. Two 1024 channel EELS spectra offset in energy by 1 eV were recorded and stored at each pixel in the 80x80 spectrum-image (25 Mbytes). An energy range of 39-89eV (20 channels/eV) are represented. During processing the spectra are either subtracted to create an artifact corrected difference spectrum, or the energy offset is numerically removed and the spectra are added to create a normal spectrum. The spectrum-images are processed into 2D floating-point images using methods and software described in [1].


Author(s):  
Delbert E. Philpott ◽  
David Leaffer

There are certain advantages for electron probe analysis if the sample can be tilted directly towards the detector. The count rate is higher, it optimizes the geometry since only one angle need be taken into account for quantitative analysis and the signal to background ratio is improved. The need for less tilt angle may be an advantage because the grid bars are not moved quite as close to each other, leaving a little more open area for observation. Our present detector (EDAX) and microscope (Philips 300) combination precludes moving the detector behind the microscope where it would point directly at the grid. Therefore, the angle of the specimen was changed in order to optimize the geometry between the specimen and the detector.


Author(s):  
Conly L. Rieder

The behavior of many cellular components, and their dynamic interactions, can be characterized in the living cell with considerable spatial and temporal resolution by video-enhanced light microscopy (video-LM). Indeed, under the appropriate conditions video-LM can be used to determine the real-time behavior of organelles ≤ 25-nm in diameter (e.g., individual microtubules—see). However, when pushed to its limit the structures and components observed within the cell by video-LM cannot be resolved nor necessarily even identified, only detected. Positive identification and a quantitative analysis often requires the corresponding electron microcopy (EM).


Author(s):  
John T. Armstrong

One of the most cited papers in the geological sciences has been that of Albee and Bence on the use of empirical " α -factors" to correct quantitative electron microprobe data. During the past 25 years this method has remained the most commonly used correction for geological samples, despite the facts that few investigators have actually determined empirical α-factors, but instead employ tables of calculated α-factors using one of the conventional "ZAF" correction programs; a number of investigators have shown that the assumption that an α-factor is constant in binary systems where there are large matrix corrections is incorrect (e.g, 2-3); and the procedure’s desirability in terms of program size and computational speed is much less important today because of developments in computing capabilities. The question thus exists whether it is time to honorably retire the Bence-Albee procedure and turn to more modern, robust correction methods. This paper proposes that, although it is perhaps time to retire the original Bence-Albee procedure, it should be replaced by a similar method based on compositiondependent polynomial α-factor expressions.


Sign in / Sign up

Export Citation Format

Share Document