Published in

Wiley, Journal of Computational Chemistry, 29(33), p. 2357-2362, 2012

DOI: 10.1002/jcc.23066

Links

Tools

Export citation

Search in Google Scholar

DataPipeline: Automated importing and fitting of large amounts of biophysical data

Journal article published in 2012 by Damien Farrell, Nielsen Je, Jens Erik Nielsen ORCID
This paper is available in a repository.
This paper is available in a repository.

Full text: Download

Green circle
Preprint: archiving allowed
Orange circle
Postprint: archiving restricted
Red circle
Published version: archiving forbidden
Data provided by SHERPA/RoMEO

Abstract

Raw data from experiments across the biological sciences comes in a large variety of text formats. In small or medium sized laboratories researchers often use an assorted collection of software to interpret, fit, and visualize their data. The spreadsheet is commonly the core component of such a workflow. The limitations of such programs for large amounts of heterogeneous data can be frustrating. We report the construction of DataPipeline, a desktop and command-line application that automates the tasks of importing, fitting, and plotting of text-based data. The software is designed to simplify the process of importing text data from various sources using simple configuration files to describe raw file formats. Once imported, curve fitting can be performed using custom fitting models designed by the user inside the application. Fitted parameters can be grouped together as new datasets to be fitted to other models and experimental uncertainties propagated to give error estimates. This software will be useful for processing of data from high through-put biological experiments or for rapid visualization of pilot data without the need for a chain of different programs to carry out each step. DataPipeline and source code is available under an open source license. The software can be freely downloaded at http://code.google.com/p/peat/downloads/list. © 2012 Wiley Periodicals, Inc.