Given a particular descriptor/method combination, some quantitative structure–activity relationship (QSAR) datasets are very predictive by random-split cross-validation while others are not. Recent ...literature in modelability suggests that the limiting issue for predictivity is in the data, not the QSAR methodology, and the limits are due to activity cliffs. Here, we investigate, on in-house data, the relative usefulness of experimental error, distribution of the activities, and activity cliff metrics in determining how predictive a dataset is likely to be. We include unmodified in-house datasets, datasets that should be perfectly predictive based only on the chemical structure, datasets where the distribution of activities is manipulated, and datasets that include a known amount of added noise. We find that activity cliff metrics determine predictivity better than the other metrics we investigated, whatever the type of dataset, consistent with the modelability literature. However, such metrics cannot distinguish real activity cliffs due to large uncertainties in the activities. We also show that a number of modern QSAR methods, and some alternative descriptors, are equally bad at predicting the activities of compounds on activity cliffs, consistent with the assumptions behind “modelability.” Finally, we relate time-split predictivity with random-split predictivity and show that different coverages of chemical space are at least as important as uncertainty in activity and/or activity cliffs in limiting predictivity.
Given the tremendous growth of bioactivity databases, the use of computational tools to predict protein targets of small molecules has been gaining importance in recent years. Applications span a ...wide range, from the ‘designed polypharmacology’ of compounds to mode-of-action analysis. In this review, we firstly survey databases that can be used for ligand-based target prediction and which have grown tremendously in size in the past. We furthermore outline methods for target prediction that exist, both based on the knowledge of bioactivities from the ligand side and methods that can be applied in situations when a protein structure is known. Applications of successful in silico target identification attempts are discussed in detail, which were based partly or in whole on computational target predictions in the first instance. This includes the authors' own experience using target prediction tools, in this case considering phenotypic antibacterial screens and the analysis of high-throughput screening data. Finally, we will conclude with the prospective application of databases to not only predict, retrospectively, the protein targets of a small molecule, but also how to design ligands with desired polypharmacology in a prospective manner.
Display omitted
► Public bioactivity databases are increasing in size at a tremendous rate. ► Integration of chemical and biological data is next key step; first examples exist. ► Target prediction and mode-of-action analysis help us rationalize compound action. ► Incorporating pathway information potentially enables multi-target drug design.
Target identification is a critical step following the discovery of small molecules that elicit a biological phenotype. The present work seeks to provide an in silico correlate of experimental target ...fishing technologies in order to rapidly fish out potential targets for compounds on the basis of chemical structure alone. A multiple-category Laplacian-modified naïve Bayesian model was trained on extended-connectivity fingerprints of compounds from 964 target classes in the WOMBAT (World Of Molecular BioAcTivity) chemogenomics database. The model was employed to predict the top three most likely protein targets for all MDDR (MDL Drug Database Report) database compounds. On average, the correct target was found 77% of the time for compounds from 10 MDDR activity classes with known targets. For MDDR compounds annotated with only therapeutic or generic activities such as “antineoplastic”, “kinase inhibitor”, or “anti-inflammatory”, the model was able to systematically deconvolute the generic activities to specific targets associated with the therapeutic effect. Examples of successful deconvolution are given, demonstrating the usefulness of the tool for improving knowledge in chemogenomics databases and for predicting new targets for orphan compounds.
High Throughput Screening (HTS) is a common approach in life sciences to discover chemical matter that modulates a biological target or phenotype. However, low assay throughput, reagents cost, or a ...flowchart that can deal with only a limited number of hits may impair screening large numbers of compounds. In this case, a subset of compounds is assayed, and in silico models are utilized to aid in iterative screening design, usually to expand around the found hits and enrich subsequent rounds for relevant chemical matter. However, this may lead to an overly narrow focus, and the diversity of compounds sampled in subsequent iterations may suffer. Active learning has been recently successfully applied in drug discovery with the goal of sampling diverse chemical space to improve model performance. Here we introduce a robust and straightforward iterative screening protocol based on naı̈ve Bayes models. Instead of following up on the compounds with the highest scores in the in silico model, we pursue compounds with very low but positive values. This includes unique chemotypes of weakly active compounds that enhance the applicability domain of the model and increase the cumulative hit rates. We show in a retrospective application to 81 Novartis assays that this protocol leads to consistently higher compound and scaffold hit rates compared to a standard expansion around hits or an active learning approach. We recommend using the weak reinforcement strategy introduced herein for iterative screening workflows.
Medicinal chemists' "intuition" is critical for success in modern drug discovery. Early in the discovery process, chemists select a subset of compounds for further research, often from many viable ...candidates. These decisions determine the success of a discovery campaign, and ultimately what kind of drugs are developed and marketed to the public. Surprisingly little is known about the cognitive aspects of chemists' decision-making when they prioritize compounds. We investigate 1) how and to what extent chemists simplify the problem of identifying promising compounds, 2) whether chemists agree with each other about the criteria used for such decisions, and 3) how accurately chemists report the criteria they use for these decisions. Chemists were surveyed and asked to select chemical fragments that they would be willing to develop into a lead compound from a set of ~4,000 available fragments. Based on each chemist's selections, computational classifiers were built to model each chemist's selection strategy. Results suggest that chemists greatly simplified the problem, typically using only 1-2 of many possible parameters when making their selections. Although chemists tended to use the same parameters to select compounds, differing value preferences for these parameters led to an overall lack of consensus in compound selections. Moreover, what little agreement there was among the chemists was largely in what fragments were undesirable. Furthermore, chemists were often unaware of the parameters (such as compound size) which were statistically significant in their selections, and overestimated the number of parameters they employed. A critical evaluation of the problem space faced by medicinal chemists and cognitive models of categorization were especially useful in understanding the low consensus between chemists.
Celotno besedilo
Dostopno za:
DOBA, IZUM, KILJ, NUK, PILJ, PNG, SAZU, SIK, UILJ, UKNU, UL, UM, UPUK
The continuing emergence of antibacterial resistance reduces the effectiveness of antibiotics and drives an ongoing search for effective replacements. Screening compound libraries for antibacterial ...activity in standard growth media has been extensively explored and may be showing diminishing returns. Inhibition of bacterial targets that are selectively important under in vivo (infection) conditions and, therefore, would be missed by conventional in vitro screens might be an alternative. Surrogate host models of infection, however, are often not suitable for high-throughput screens. Here, we adapted a medium-throughput Tetrahymena pyriformis surrogate host model that was successfully used to identify inhibitors of a hyperviscous Klebsiella pneumoniae strain to a high-throughput format and screened circa 1.2 million compounds. The screen was robust and identified confirmed hits from different chemical classes with potent inhibition of K. pneumoniae growth in the presence of T. pyriformis that lacked any appreciable direct antibacterial activity. Several of these appeared to inhibit capsule/mucoidy, which are key virulence factors in hypervirulent K. pneumoniae. A weakly antibacterial inhibitor of LpxC (essential for the synthesis of the lipid A moiety of lipopolysaccharides) also appeared to be more active in the presence of T. pyriformis, which is consistent with the role of LPS in virulence as well as viability in K. pneumoniae.
A first step in fragment-based drug discovery (FBDD) often entails a fragment-based screen (FBS) to identify fragment “hits.” However, the integration of conflicting results from orthogonal screens ...remains a challenge. Here we present a meta-analysis of 35 fragment-based campaigns at Novartis, which employed a generic 1400-fragment library against diverse target families using various biophysical and biochemical techniques. By statistically interrogating the multidimensional FBS data, we sought to investigate three questions: (1) What makes a fragment amenable for FBS? (2) How do hits from different fragment screening technologies and target classes compare with each other? (3) What is the best way to pair FBS assay technologies? In doing so, we identified substructures that were privileged for specific target classes, as well as fragments that were privileged for authentic activity against many targets. We also revealed some of the discrepancies between technologies. Finally, we uncovered a simple rule of thumb in screening strategy: when choosing two technologies for a campaign, pairing a biochemical and biophysical screen tends to yield the greatest coverage of authentic hits.
The off-rate (koff) of the T cell receptor (TCR)/peptide-major histocompatibility complex class I (pMHCI) interaction, and hence its half-life, is the principal kinetic feature that determines the ...biological outcome of TCR ligation. However, it is unclear whether the CD8 coreceptor, which binds pMHCI at a distinct site, influences this parameter. Although biophysical studies with soluble proteins show that TCR and CD8 do not bind cooperatively to pMHCI, accumulating evidence suggests that TCR associates with CD8 on the T cell surface. Here, we titrated and quantified the contribution of CD8 to TCR/pMHCI dissociation in membrane-constrained interactions using a panel of engineered pMHCI mutants that retain faithful TCR interactions but exhibit a spectrum of affinities for CD8 of >1,000-fold. Data modeling generates a “stabilization factor” that preferentially increases the predicted TCR triggering rate for low affinity pMHCI ligands, thereby suggesting an important role for CD8 in the phenomenon of T cell cross-reactivity.
Lead discovery in the pharmaceutical environment is largely an industrial-scale process in which it is typical to screen 1–5 million compounds in a matter of weeks using High Throughput Screening ...(HTS). This process is a very costly endeavor. Typically a HTS campaign of 1 million compounds will cost anywhere from $500
000 to $1
000
000. There is consequently a great deal of pressure to maximize the return on investment by finding fast and more effective ways to screen. A panacea that has emerged over the past few years to help address this issue is
in silico screening.
In silico screening is now incorporated in all areas of lead discovery; from target identification and library design, to hit analysis and compound profiling. However, as lead discovery has evolved over the past few years, so has the role of
in silico screening.
CD8+ cytotoxic T lymphocytes (CTL) are key determinants of immunity to intracellular pathogens and neoplastic cells. Recognition of specific antigens in the form of peptide‐MHC class I complexes ...(pMHCI) presented on the target cell surface is mediated by T cell receptor (TCR) engagement. The CD8 coreceptor binds to invariant domains of pMHCI and facilitates antigen recognition. Here, we investigate the biological effects of a Q115E substitution in the α2 domain of human leukocyte antigen (HLA)‐A*0201 that enhances CD8 binding by ∼50% without altering TCR/pMHCI interactions. Soluble and cell surface‐expressed forms of Q115E HLA‐A*0201 exhibit enhanced recognition by CTL without loss of specificity. These CD8‐enhanced antigens induce greater CD3 ζ chain phosphorylation in cognate CTL leading to substantial increases in cytokine production, proliferation and priming of naive T cells. This effect provides a fundamental new mechanism with which to enhance cellular immunity to specific T cell antigens.