Extensive Combinatorial Refinement process creates computerized molecular models more automatically

It's difficult and time-consuming to produce accurate computer models of molecules, primarily because traditional modeling methods are limited in their ability to handle alternative molecular shapes and, consequently, are subject to multiple errors.

Moreover, the traditional approach uses mathematical formulas or algorithms that are run sequentially, refining the structural details of the model with each separate algorithm-a method that has been revolutionized by personal computing, but still requires labor-intensive human intervention for error correction.

A new method developed by scientists on the Florida campus of The Scripps Research Institute (TSRI) takes another tack entirely, combining existing formulas in a kind of algorithmic stew to gain a better picture of molecular structural diversity that is then used to eliminate errors and improve the final model.

The method was described in a paper published online ahead of print on September 26, 2013 by the journal Structure.

The new process, called Extensive Combinatorial Refinement (ExCoR), could help improve the development of drug candidates that depend to a great degree on detailed structural analysis to determine how they work against specific disease targets.

"Our combinatorial method creates computerized molecular models in a more automated way," said Kendall Nettles, a TSRI associate professor who led the study. "This is an important component of drug discovery-to do them in a more automated fashion will significantly help the process."

Improvement and Some Surprises

In the study, the scientists subjected more than 50 molecular structures to 256 distinct combinations of algorithms and refinement factors that eventually totaled more than 12,000 independent refinement runs.

Nettles and his colleagues measured the improvement in the models by what is known as the R-factor, which measures the similarity between the actual structure of the molecule and the experimental model-in other words, just how closely the refined structure model can predict the factual data.

"Lowering that R-factor is the goal-that's the selection process for finding the best algorithms," Nettles said.

While the study found that no single algorithm consistently produced the best model, the scientists did find some surprises.

"Some algorithms, if you combine them, tend to work better at producing a refined model," said Research Associate Jerome C. Nwachukwu, the first author of the study. "What we didn't expect was two algorithms that worked separately but didn't work in combination."

It is this strange overlap makes it impossible to predict which combinations of algorithms will work best for an individual structure.

"The refinement effects of the various algorithms depend on the structure itself," Nwachukwu said.

Comments

The opinions expressed here are the views of the writer and do not necessarily reflect the views and opinions of News Medical.
Post a new comment
Post

While we only use edited and approved content for Azthena answers, it may on occasions provide incorrect responses. Please confirm any data provided with the related suppliers or authors. We do not provide medical advice, if you search for medical information you must always consult a medical professional before acting on any information provided.

Your questions, but not your email details will be shared with OpenAI and retained for 30 days in accordance with their privacy principles.

Please do not ask questions that use sensitive or confidential information.

Read the full Terms & Conditions.

You might also like...
Research links COVID-19 vaccines to temporary facial palsy in over 5,000 patients