Hi All!

I have some experience with R, but less experience writing scripts using 
R and have run into a challenge that I hope someone can help me with.

  I have multiple .csv files of data each with the same 3 columns of 
data, but potentially of varying lengths (some data files are from short 
measurements, others from longer ones). One file for example might look 
like this...

Time, O2_conc, Chla_conc

0,270,300

10, 260, 280

20, 245, 268

30, 233, 238

40, 222, 212

50, 215, 201

60, 208, 193

70, 206, 191

80, 207,189

90, 206, 186

100, 206, 183

110, 207, 178

120, 205, 174

130, 240, 171

140, 270, 155

I am looking for an efficient means of batch (or sequentially) 
processing these files so that I can
1. import each data file

2. find the minimum value recorded in column 2 and the previous 5 data 
points

3. and average these 10 values to get a mean, minimum value.

Currently I have imported the data files using the following

filenames=list.files()

library(plyr)

import.list=adply(filenames, 1, read.csv)

and I know how to write a code to calculate the minimum value and the 5 
preceding values in a single column, in a single file. I think the 
problem I am running into is scaling this code up so that I can import 
multiple files and calculating mean, minimum value for the 2^nd column 
in each of them.

Can anyone offer some advice on how to batch processes a whole bunch of 
files? I need to load them in, but then analyze them too.

Thank you so much,

Nate


        [[alternative HTML version deleted]]

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to