Mathematical and computational approaches to epidemic modeling: a comprehensive review
Mathematical and computational approaches are important tools for understanding epidemic spread patterns and evaluating policies of disease control. In recent years, epidemiology has become increasingly integrated with mathematics, sociology, management science, complexity science, and computer science. The cross of multiple disciplines has caused rapid development of mathematical and computational approaches to epidemic modeling. In this article, we carry out a comprehensive review of epidemic models to provide an insight into the literature of epidemic modeling and simulation. We introduce major epidemic models in three directions, including mathematical models, complex network models, and agent-based models. We discuss the principles, applications, advantages, and limitations of these models. Meanwhile, we also propose some future research directions in epidemic modeling.
Social influence and spread dynamics in social networks
Social networks often serve as a critical medium for information dissemination, diffusion of epidemics, and spread of behavior, by shared activities or similarities between individuals. Recently, we have witnessed an explosion of interest in studying social influence and spread dynamics in social networks. To date, relatively little material has been provided on a comprehensive review in this field. This brief survey addresses this issue. We present the current significant empirical studies on real social systems, including network construction methods, measures of network, and newly empirical results. We then provide a concise description of some related social models from both macro- and micro-level perspectives. Due to the difficulties in combining real data and simulation data for verifying and validating real social systems, we further emphasize the current research results of computational experiments. We hope this paper can provide researchers significant insights into better understanding the characteristics of personal influence and spread patterns in large-scale social systems.
Image smog restoration using oblique gradient profile prior and energy minimization
Removing the smog from digital images is a challenging pre-processing tool in various imaging systems. Therefore, many smog removal (i.e., desmogging) models are proposed so far to remove the effect of smog from images. The desmogging models are based upon a physical model, it means it requires efficient estimation of transmission map and atmospheric veil from a single smoggy image. Therefore, many prior based restoration models are proposed in the literature to estimate the transmission map and an atmospheric veil. However, these models utilized computationally extensive minimization of an energy function. Also, the existing restoration models suffer from various issues such as distortion of texture, edges, and colors. Therefore, in this paper, a convolutional neural network (CNN) is used to estimate the physical attributes of smoggy images. Oblique gradient channel prior (OGCP) is utilized to restore the smoggy images. Initially, a dataset of smoggy and sunny images are obtained. Thereafter, we have trained CNN to estimate the smog gradient from smoggy images. Finally, based upon the computed smog gradient, OGCP is utilized to restore the still smoggy images. Performance analyses reveal that the proposed CNN-OGCP based desmogging model outperforms the existing desmogging models in terms of various performance metrics.
Challenges and future directions of secure federated learning: a survey
Federated learning came into being with the increasing concern of privacy security, as people's sensitive information is being exposed under the era of big data. It is an algorithm that does not collect users' raw data, but aggregates model parameters from each client and therefore protects user's privacy. Nonetheless, due to the inherent distributed nature of federated learning, it is more vulnerable under attacks since users may upload malicious data to break down the federated learning server. In addition, some recent studies have shown that attackers can recover information merely from parameters. Hence, there is still lots of room to improve the current federated learning frameworks. In this survey, we give a brief review of the state-of-the-art federated learning techniques and detailedly discuss the improvement of federated learning. Several open issues and existing solutions in federated learning are discussed. We also point out the future research directions of federated learning.
A privacy-preserving group encryption scheme with identity exposure
AE-TPGG: a novel autoencoder-based approach for single-cell RNA-seq data imputation and dimensionality reduction
Single-cell RNA sequencing (scRNA-seq) technology has become an effective tool for high-throughout transcriptomic study, which circumvents the averaging artifacts corresponding to bulk RNA-seq technology, yielding new perspectives on the cellular diversity of potential superficially homogeneous populations. Although various sequencing techniques have decreased the amplification bias and improved capture efficiency caused by the low amount of starting material, the technical noise and biological variation are inevitably introduced into experimental process, resulting in high dropout events, which greatly hinder the downstream analysis. Considering the bimodal expression pattern and the right-skewed characteristic existed in normalized scRNA-seq data, we propose a customized autoencoder based on a two-part-generalized-gamma distribution (AE-TPGG) for scRNA-seq data analysis, which takes mixed discrete-continuous random variables of scRNA-seq data into account using a two-part model and utilizes the generalized gamma (GG) distribution, for fitting the positive and right-skewed continuous data. The adopted autoencoder enables AE-TPGG to captures the inherent relationship between genes. In addition to the ability of achieving low-dimensional representation, the AE-TPGG model also provides a denoised imputation according to statistical characteristic of gene expression. Results on real datasets demonstrate that our proposed model is competitive to current imputation methods and ameliorates a diverse set of typical scRNA-seq data analyses.
FragDPI: a novel drug-protein interaction prediction model based on fragment understanding and unified coding
Prediction of drug-protein binding is critical for virtual drug screening. Many deep learning methods have been proposed to predict the drug-protein binding based on protein sequences and drug representation sequences. However, most existing methods extract features from protein and drug sequences separately. As a result, they can not learn the features characterizing the drug-protein interactions. In addition, the existing methods encode the protein (drug) sequence usually based on the assumption that each amino acid (atom) has the same contribution to the binding, ignoring different impacts of different amino acids (atoms) on the binding. However, the event of drug-protein binding usually occurs between conserved residue fragments in the protein sequence and atom fragments of the drug molecule. Therefore, a more comprehensive encoding strategy is required to extract information from the conserved fragments. In this paper, we propose a novel model, named FragDPI, to predict the drug-protein binding affinity. Unlike other methods, we encode the sequences based on the conserved fragments and encode the protein and drug into a unified vector. Moreover, we adopt a novel two-step training strategy to train FragDPI. The pre-training step is to learn the interactions between different fragments using unsupervised learning. The fine-tuning step is for predicting the binding affinities using supervised learning. The experiment results have illustrated the superiority of FragDPI.
