scholarly journals High Efficiency Spam Filtering: A Manifold Learning-Based Approach

2021 ◽  
Vol 2021 ◽  
pp. 1-7
Author(s):  
Chao Wang ◽  
Qun Li ◽  
Tian-yu Ren ◽  
Xiao-hu Wang ◽  
Guang-xin Guo

Spam filtering, which refers to detecting unsolicited, unwanted, and virus-infested emails, is a significant problem because spam emails lead to unnecessary costs of Internet resources, waste of people’s time, and even loss of property. Support vector machine (SVM) is the state-of-the-art method for high accuracy spam filtering. However, SVM incurs high time complexity because of the high dimensionality of the emails. In this study, we propose a manifold learning-based approach for time-efficient spam filtering. From the experiments that most of the features are not decisive, we can obtain the viewpoint that only a minor part of the spam emails can be detected using the nondecisive features. Based on the insight, we propose to employ the Laplace feature map algorithm to obtain the geometrical information from the email text datasets and extract the decisive features. Then, the extracted features are used as the input of SVM to spam filtering. We conduct extensive experiments on three datasets, and the evaluation results indicate the high accuracy time efficiency of our proposed algorithm.

2021 ◽  
Author(s):  
Nolan Grieves ◽  
François Bouchy ◽  
René Doyon ◽  
Etienne Artigau ◽  
Lison Malo ◽  
...  

<p>The Near-InfraRed Planet Searcher (NIRPS) is designed to be an ultra-stable infrared spectrograph to be installed on ESO’s 3.6 m Telescope in La Silla, Chile. NIRPS is an adaptive optics (AO) fiber-fed spectrograph operating from 0.98 to 1.8 μm and will be operated simultaneously with the optical high-resolution spectrograph HARPS. NIRPS can operate in two modes fed by two different fiber links permanently mounted at the Cassegrain focus that use either 0.4 arcsecond-fibers for the High Accuracy Mode (HAM) or 0.9 arcsecond-fibers for the High Efficiency Mode (HEM). The wavelength range of NIRPS is optimal for low-mass M dwarfs and the simultaneous NIRPS and HARPS observations will improve stellar activity filtering methods given their different wavelength coverages. The NIRPS front-end and AO system were already tested on-sky at La Silla. The spectrograph and back-end is being shipped to La Silla and installed in Summer/Fall 2021. Already we have adapted the state-of-the-art ESPRESSO data reduction pipeline for NIRPS, obtained accurate wavelength solutions with a Uranium Neon lamp, and obtained drift stability results below 50 cm/s with a Fabry–Pérot etalon. We discuss the current and expected instrument performance and the expected results of NIRPS.</p>


2007 ◽  
Vol 11 (4) ◽  
pp. 5-40 ◽  
Author(s):  
Bo Leckner

Co-combustion of biomass or waste together with a base fuel in a boiler is a simple and economically suitable way to replace fossil fuels by biomass and to utilize waste. Co-combustion in a high-efficiency power station means utilization of biomass and waste with a higher thermal efficiency than what otherwise had been possible. Due to transport limitations, the additional fuel will only supply a minor part (less than a few hundreds MW fuel) of the energy in a plant. There are several options: co-combustion with coal in pulverized or fluidized bed boilers, combustion on added grates inserted in pulverized coal boilers, combustors for added fuel coupled in parallel to the steam circuit of a power plant, external gas producers delivering its gas to replace an oil, gas or pulverized fuel burner. Furthermore biomass can be used for reburning in order to reduce NO emissions or for afterburning to reduce N2O emissions in fluidized bed boilers. Combination of fuels can give rise to positive or negative synergy effects, of which the best known are the interactions between S, Cl, K, Al, and Si that may give rise to or prevent deposits on tubes or on catalyst surfaces, or that may have an influence on the formation of dioxins. With better knowledge of these effects the positive ones can be utilized and the negative ones can be avoided.


Symmetry ◽  
2019 ◽  
Vol 11 (11) ◽  
pp. 1343 ◽  
Author(s):  
Zhenzhen Zhang ◽  
Changbo Liu ◽  
Zhaohong Li ◽  
Lifang Yu ◽  
Huanma Yan

High Efficiency Video Coding (HEVC) is a worldwide popular video coding standard due to its high coding efficiency. To make profits, forgers prefer to transcode videos from previous standards such as H.264/AVC to HEVC. To deal with this issue, an efficient method is proposed to expose such transcoded HEVC videos based on coding unit (CU) and prediction unit (PU) partition types. CU and PU partitioning are two unique syntactic units of HEVC that can reflect a video’s compression history. In this paper, CU and PU partition types of I pictures and P pictures are firstly extracted. Then, their mean frequencies are calculated and concatenated as distinguishing features, which are further sent to a support vector machine (SVM) for classification. Experimental results show that the proposed method can identify transcoded HEVC videos with high accuracy and has strong robustness against frame-deletion and shifted Group of Pictures (GOP) structure attacks.


2021 ◽  
Vol 37 (1--4) ◽  
pp. 1-15
Author(s):  
Chaim Baskin ◽  
Natan Liss ◽  
Eli Schwartz ◽  
Evgenii Zheltonozhskii ◽  
Raja Giryes ◽  
...  

We present a novel method for neural network quantization. Our method, named UNIQ , emulates a non-uniform k -quantile quantizer and adapts the model to perform well with quantized weights by injecting noise to the weights at training time. As a by-product of injecting noise to weights, we find that activations can also be quantized to as low as 8-bit with only a minor accuracy degradation. Our non-uniform quantization approach provides a novel alternative to the existing uniform quantization techniques for neural networks. We further propose a novel complexity metric of number of bit operations performed (BOPs), and we show that this metric has a linear relation with logic utilization and power. We suggest evaluating the trade-off of accuracy vs. complexity (BOPs). The proposed method, when evaluated on ResNet18/34/50 and MobileNet on ImageNet, outperforms the prior state of the art both in the low-complexity regime and the high accuracy regime. We demonstrate the practical applicability of this approach, by implementing our non-uniformly quantized CNN on FPGA.


Author(s):  
Lidiya Derbenyova

The article explores the role of antropoetonyms in the reader’s “horizon of expectation” formation. As a kind of “text in the text”, antropoetonyms are concentrating a large amount of information on a minor part of the text, reflecting the main theme of the work. As a “text” this class of poetonyms performs a number of functions: transmission and storage of information, generation of new meanings, the function of “cultural memory”, which explains the readers’ “horizon of expectations”. In analyzing the context of the literary work we should consider the function of antropoetonyms in vertical context (the link between artistic and other texts, and the groundwork system of culture), as well as in the context of the horizontal one (times’ connection realized in the communication chain from the word to the text; the author’s intention). In this aspect, the role of antropoetonyms in the structure of the literary text is extremely significant because antropoetonyms convey an associative nature, generating a complex mechanism of allusions. It’s an open fact that they always transmit information about the preceding text and suggest a double decoding. On the one hand, the recipient decodes this information, on the other – accepts this as a sort of hidden, “secret” sense.


2020 ◽  
Vol 17 (6) ◽  
pp. 847-856
Author(s):  
Shengbing Ren ◽  
Xiang Zhang

The problem of synthesizing adequate inductive invariants lies at the heart of automated software verification. The state-of-the-art machine learning algorithms for synthesizing invariants have gradually shown its excellent performance. However, synthesizing disjunctive invariants is a difficult task. In this paper, we propose a method k++ Support Vector Machine (SVM) integrating k-means++ and SVM to synthesize conjunctive and disjunctive invariants. At first, given a program, we start with executing the program to collect program states. Next, k++SVM adopts k-means++ to cluster the positive samples and then applies SVM to distinguish each positive sample cluster from all negative samples to synthesize the candidate invariants. Finally, a set of theories founded on Hoare logic are adopted to check whether the candidate invariants are true invariants. If the candidate invariants fail the check, we should sample more states and repeat our algorithm. The experimental results show that k++SVM is compatible with the algorithms for Intersection Of Half-space (IOH) and more efficient than the tool of Interproc. Furthermore, it is shown that our method can synthesize conjunctive and disjunctive invariants automatically


1990 ◽  
Vol 55 (1) ◽  
pp. 1-9 ◽  
Author(s):  
Zdeněk Friedl ◽  
Stanislav Böhm

The relative enthalpies of proton transfer δ ΔH0and homolytic bond strengths δDH0(B-H+) were calculated by the MNDO method for the sp and ap conformers of 4-flurobutylamine. The data obtained, along with the experimental gas phase basicities, are compared with the values predicted by the electrostatic theory. It is shown that the substituent polar effects FD on the basicities of amines are predominantly due to interactions in their protonated forms (X-B-H+) and/or radical-cations (X-B+.), those in the neutral species (X-B) playing a minor part. A contribution, which is considerably more significant in the sp conformer than in the ap conformer, arises probably also from substituent effects on the homolytic bond strength DH0(B-H+.


Nanomaterials ◽  
2021 ◽  
Vol 11 (3) ◽  
pp. 560
Author(s):  
Alexandra Carvalho ◽  
Mariana C. F. Costa ◽  
Valeria S. Marangoni ◽  
Pei Rou Ng ◽  
Thi Le Hang Nguyen ◽  
...  

We show that the degree of oxidation of graphene oxide (GO) can be obtained by using a combination of state-of-the-art ab initio computational modeling and X-ray photoemission spectroscopy (XPS). We show that the shift of the XPS C1s peak relative to pristine graphene, ΔEC1s, can be described with high accuracy by ΔEC1s=A(cO−cl)2+E0, where c0 is the oxygen concentration, A=52.3 eV, cl=0.122, and E0=1.22 eV. Our results demonstrate a precise determination of the oxygen content of GO samples.


Data ◽  
2021 ◽  
Vol 6 (8) ◽  
pp. 87
Author(s):  
Sara Ferreira ◽  
Mário Antunes ◽  
Manuel E. Correia

Deepfake and manipulated digital photos and videos are being increasingly used in a myriad of cybercrimes. Ransomware, the dissemination of fake news, and digital kidnapping-related crimes are the most recurrent, in which tampered multimedia content has been the primordial disseminating vehicle. Digital forensic analysis tools are being widely used by criminal investigations to automate the identification of digital evidence in seized electronic equipment. The number of files to be processed and the complexity of the crimes under analysis have highlighted the need to employ efficient digital forensics techniques grounded on state-of-the-art technologies. Machine Learning (ML) researchers have been challenged to apply techniques and methods to improve the automatic detection of manipulated multimedia content. However, the implementation of such methods have not yet been massively incorporated into digital forensic tools, mostly due to the lack of realistic and well-structured datasets of photos and videos. The diversity and richness of the datasets are crucial to benchmark the ML models and to evaluate their appropriateness to be applied in real-world digital forensics applications. An example is the development of third-party modules for the widely used Autopsy digital forensic application. This paper presents a dataset obtained by extracting a set of simple features from genuine and manipulated photos and videos, which are part of state-of-the-art existing datasets. The resulting dataset is balanced, and each entry comprises a label and a vector of numeric values corresponding to the features extracted through a Discrete Fourier Transform (DFT). The dataset is available in a GitHub repository, and the total amount of photos and video frames is 40,588 and 12,400, respectively. The dataset was validated and benchmarked with deep learning Convolutional Neural Networks (CNN) and Support Vector Machines (SVM) methods; however, a plethora of other existing ones can be applied. Generically, the results show a better F1-score for CNN when comparing with SVM, both for photos and videos processing. CNN achieved an F1-score of 0.9968 and 0.8415 for photos and videos, respectively. Regarding SVM, the results obtained with 5-fold cross-validation are 0.9953 and 0.7955, respectively, for photos and videos processing. A set of methods written in Python is available for the researchers, namely to preprocess and extract the features from the original photos and videos files and to build the training and testing sets. Additional methods are also available to convert the original PKL files into CSV and TXT, which gives more flexibility for the ML researchers to use the dataset on existing ML frameworks and tools.


Sign in / Sign up

Export Citation Format

Share Document