I have a large (100MB, 940k lines) csv file. If I run
g <- read.csv('combineTfe-dataFrame-TD.dat')
write.csv(g, file="tmp.data");
Now ?mean returns without printing anything to screen and
system("touch test.dat") does nothing.
If instead I just write out the top of the data, as in
g <- read.cs
On Fri, Aug 22, 2008 at 11:28 AM, Todd Hatfield <[EMAIL PROTECTED]> wrote:
> I would like to control the inner margins of a lattice graph. The graph is
> a single superposed panel, but there is too much white space around the data
> for my liking. I've played around with some of the layout option
Try 'check.names=FALSE' in the read.table call.
On Fri, Aug 22, 2008 at 6:11 PM, Sébastien <[EMAIL PROTECTED]> wrote:
> Dear R users,
>
> I have to import some csv files in which column headers contain the
> character "~". Following the following import call, the character seems to
> be replaced b
rownames(yourMatrix) <- sprintf("%.1f", seq(.1, 7, .1))
On Fri, Aug 22, 2008 at 6:48 PM, Jan Akko Eleveld
<[EMAIL PROTECTED]> wrote:
>
> Question:
>
> I created an array with 164 columns and 70 rows.
>
> I would like every column to have unit value 1
> and the rows 0.1, 0.2, 0.3 up to 7. How can I
Here is one way of doing it:
> lt
$a
[1] 1 2 3
$b
[1] NA
$c
[1] "a" "b" "c"
$d
[1] NA
> lt[!unlist(sapply(lt, function(x) any(is.na(x[1]]
$a
[1] 1 2 3
$c
[1] "a" "b" "c"
>
On Fri, Aug 22, 2008 at 2:41 PM, Dong-hyun Oh <[EMAIL PROTECTED]> wrote:
> Dear useRs,
>
> I would like to know th
Try 'check.names=FALSE' in the read.table call.
On Fri, Aug 22, 2008 at 6:02 PM, kayj <[EMAIL PROTECTED]> wrote:
>
>
> I am trying to use rbind to have the two data on the top of each other but I
> am getting an extra X on the column header and the rows are numberd , How to
> get rid of this probl
WTF mate. you think im about wasting time. I been seeking a solution for
this for two weeks. excuse my lack of programming savvy, but this is why i
posted.
bartjoosen wrote:
>
> Please take a look at the manuals before posting (DO read the posting
> guide!)
>
> Bart
> PS: yourdataframe[yourda
I am trying to use rbind to have the two data on the top of each other but I
am getting an extra X on the column header and the rows are numberd , How to
get rid of this problem? I appreciate your help
x1<- read.table(file="data1.txt", header=T, sep="\t")
x2<-read.table(file="data2.txt", head
Hi my name is leidy. I am a student of sistem ingenier, i do not speak
english.
Now I am working in a proyect using R but with a interface of java, I have a
problem because a need acces to p
> s<-box.cox.powers(resp,2)
> s
Box-Cox Transformation to Normality
Est.Power Std.Err. Wald(Power
Dear useRs,
I would like to know the way of deleting NA in list().
Following is a example.
lt <- list(a = 1:3, b = NA, c = letters[1:3], d = NA)
for(i in length(lt):1) {
if(is.na(lt[[i]])) lt[[i]] <- NULL
}
How to simplify for() loop by using (l)apply family?
Thank you in advance.
Le vendredi 22 août 2008 à 15:16 -0400, John Tillinghast a écrit :
> I'm trying to figure this out with "Writing R Extensions" but there's not a
> lot of detail on this issue.
> I want to write a (very simple really) C external that will be able to take
> "..." as an argument.
> (It's for optimizin
Jakob,
Could you please tell me how you got your 2-mode data INTO R/statnet/sna,
etc... ?
I can do many things with this data in ucinet, and even export it to pajek,
but I cannot get it to import even using the read.paj routine. I have about
620 actors involved in 230 events. When I convert it
Hi, when I try to run GLS, R crashes. I have seen on the mailing list that
somebody has encountered the problem before (
http://tolstoy.newcastle.edu.au/R/help/05/01/10915.html ), but there seemed
to be no solution / the thread appears incomplete. The program runs gls1
(without correlation structur
Question:
I created an array with 164 columns and 70 rows.
I would like every column to have unit value 1
and the rows 0.1, 0.2, 0.3 up to 7. How can I define that?
Thanks for your help!
Good weekend!
Akko
_
[[a
Dear R users,
I have to import some csv files in which column headers contain the
character "~". Following the following import call, the character seems
to be replaced by dots in the column names of my data frame. Plus, I
cannot query names(mydata) to find the column index which header should
I have been experimenting with interactive packages such iplots and playwith.
Consider the following sample dataset:
A B C D
1 5 5 9
3 2 8 4
1 7 3 0
7 2 2 6
Let's say I make a plot of variable A. I would like to be able to click on
a data point (e.g. 3) and have a pop-up window
On 8/22/2008 5:34 PM, Ranney, Steven wrote:
> All -
>
> Not sure if this is a real programming question, but here goes:
>
> I have data that looks like
>
> Lake Length Weight
> 1 158 45
> 1 179 70
> 1 200 125
> 1 202 150
> 1 206 145
> 1 209 165
"Sun, Ying" asked:
I have a genotype data for both case and controls and would
like to calculate the HW p-value. However, since the
number of one genotype is 0, I got wierd result. Would
someone help me to figure it out? Or confirm it's right?
Thanks a lot.
...
HWE.exact(g1)
Exact Test
All -
Not sure if this is a real programming question, but here goes:
I have data that looks like
LakeLength Weight
1 158 45
1 179 70
1 200 125
1 202 150
1 206 145
1 209 165
1 210 140
1 215 175
1 216
Hi Hadley,
on 08/22/2008 03:52 PM hadley wickham wrote:
> Dear all,
>
> I'm having problems installing the colorspace package on Red Hat
> Enterprise Linux 5:
>
> * Installing *source* package 'colorspace' ...
> ** libs
> gcc -std=gnu99 -I/usr/lib/R/include -I/usr/lib/R/include
> -I/usr/local/in
Dear all,
I'm having problems installing the colorspace package on Red Hat
Enterprise Linux 5:
* Installing *source* package 'colorspace' ...
** libs
gcc -std=gnu99 -I/usr/lib/R/include -I/usr/lib/R/include
-I/usr/local/include-fpic -O2 -g -pipe -Wall
-Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fs
Bernhard and Irina,
While tailoring the Rcmdr menu directly does work, I strongly do not
recommend it. It will lead to mass confusion, most likely for the author,
several
months from now when it is installed on a different machine or when R or
Rcmdr is updated.
It is much cleaner to have a sepa
We have different starting points. Please be sure that your modularity
allows a cleaned region as well as a history log to be the input to your
next
step. The history log is incomplete; lines sent to the *R* buffer by C-c
C-n are
explicitly excluded from history. Lines picked up from a saved tra
Here's one approach:
try_default <- function(expr, default = NA, quiet = FALSE) {
result <- default
if (quiet) {
tryCatch(result <- expr, error = function(e) {})
} else {
try(result <- expr)
}
result
}
failwith <- function(default = NULL, f, quiet = FALSE) {
function(...) try_
Hi All:
This was discussed in the R-developers list (see the thread
starting at http://tolstoy.newcastle.edu.au/R/e3/devel/
07/12/0560.html). It has to do with the behavior of sd() when the
entire vector is NA. The behavior has changed between 2.6 and 2.7.1
as follows:
Run in Version
Great, thanks for your reply!
I've just tracked down a copy of Pinheiro & Bates (2000) so I'll look at
that, too.
Thanks again, Mark
On Fri, Aug 22, 2008 at 9:56 AM, Christoph Scherber <
[EMAIL PROTECTED]> wrote:
> Dear Mark,
>
> I would include the repeated measure as the smallest stratum in
I'm trying to figure this out with "Writing R Extensions" but there's not a
lot of detail on this issue.
I want to write a (very simple really) C external that will be able to take
"..." as an argument.
(It's for optimizing a function that may have several parameters besides the
ones being optimize
There are some interface commands in the TeachingDemos package (see ?gp.open)
for communicating between R and Gnuplot. These commands are pretty basic, but
looking at the code may give you some ideas of what to do next in your project
(or maybe what not to do).
Hope this helps,
--
Gregory (Gr
Hello,
I am trying to send commands to GNUPLOT to load a .plt file that was generated
with a C++ software module called with R, and to replot.
I am able to open the program with shell.exec (below) but I am at a loss at
what do next. Any suggestions??
shell.exec("C:\\ gnuplot \\ bin \\ wgnu
I would like to control the inner margins of a lattice graph. The graph is
a single superposed panel, but there is too much white space around the data
for my liking. I've played around with some of the layout options, but so
far have found only how to control outer margins.
R version 2.7.1 (200
You don't need to test that the _sample_ variances are different: They
already are. Statistical tests of hypotheses are not about sample
statistics, but distributional charateristics.
It seems to me that reinforcement of some basic stat concept may do you
quite a bit of good. If you don't have
Have you read the documentation to either of the functions you are using?
?bartlett.test
"Performs Bartlett's test of the null that the variances in each of the
groups (samples) are the same."
This explicitly tells you what is being tested, i.e. the null tested is that
var1 = var2.
?rnorm
G
Hi Brian,
Am 22.08.2008 um 19:22 schrieb Prof Brian Ripley:
On Fri, 22 Aug 2008, Martin Ballaschk wrote:
My solution is reading the files without the headers (skip = 1) and
seperately reading the headers with scan (scan("myfile.CSV", what =
"character", sep = "\t", nlines = 1). After throw
On the general question on how to create a dataset that matches the
frequencies in a table, function as.data.frame can be useful. It takes as
argument an object of a class 'table' and returns a data frame of
frequencies.
Consider for example table 6.1 of Fleiss et al (3rd Ed):
> birth.weight <-
On Fri, 22 Aug 2008, Martin Ballaschk wrote:
Hi folks,
thank you for your friendly and immediate help!
Am 22.08.2008 um 17:14 schrieb Prof Brian Ripley:
Or, better, use header=FALSE, skip=1 and the col.names arg of read.table().
My solution is reading the files without the headers (skip = 1
Try this:
> myvector<-c("ajkhfkiuwe","Variable(kg/min)")
> gsub( "\\(kg/min\\)", "va", myvector )
Does that come close to what you want? If not, maybe an example of what you
want the result to look like,
--
Gregory (Greg) L. Snow Ph.D.
Statistical Data Center
Intermountain Healthcare
[EMAIL PR
Hi folks,
thank you for your friendly and immediate help!
Am 22.08.2008 um 17:14 schrieb Prof Brian Ripley:
Or, better, use header=FALSE, skip=1 and the col.names arg of
read.table().
My solution is reading the files without the headers (skip = 1) and
seperately reading the headers with sc
artimon wrote:
Dear Users,
I am new to both of things, so do not blame me too much...
I am busy with semiparametric regression and use WinBUGS to sample
posteriors.
The code to call Winbugs is as follows:
data<- list("y","X","n","m") #My variables
inits.beta <- rep(0,K)
ini
I *think* I'm making progress, but I'm still failing at the same step. My rake
call fails with:
Error in postStratify.survey.design(design, strata[[i]],
population.margins[[i]], :
Stratifying variables don't match
To my naïve eyes, it seems that my factors are "in the wrong order". If so,
Hello,
I have an expression that I want to substitute for something else.
myvector<-c("ajkhfkiuwe","Variable(kg/min)")
if I use the following code to extract "variable(kg/min)" and substitute it for
"va"
gsub(myvector[grep("var", myvector, ignore=T)], "va", myvector)
grep identifies the e
Just as a follow-up on T. Lumley's post, 2 citations that may be useful in
reference to application of quantile regression with survey samples are:
Bassett and Saleh. 1994. L_1 estimation of the median of a survey
population. Nonparametric Statistics 3: 277-283. (L_1 estimation of
median is
Here's a function that does what you asked, you may need to adjust the column
names of your data frames before using it. If all your data frames are
similar (same number of rows, same years) then try do.call('cbind',
yourList).
#This function takes a list of data frames and merges them into one
Dear Users,
I am new to both of things, so do not blame me too much...
I am busy with semiparametric regression and use WinBUGS to sample
posteriors.
The code to call Winbugs is as follows:
data<- list("y","X","n","m") #My variables
inits.beta <- rep(0,K)
inits.beta0 <- 0
i
Thank you, thank you everyone!
This last worked the best for what I need to do. The other options were
definitely helpful in that they help me understand how R is structured
(something I haven't been able to get my head around yet) and gave me good
ideas about how to manipulate data, however
I don't think that the election data is the right data to demonstrate Kappa,
you need subjects that are classified by 2 or more different raters/methods.
The election data could be considered classifying the voters into which party
they voted for, but you only have 1 rater. Maybe if you had so
Try this: it came from the help on 'bxp' which is called from boxplot:
boxplot(count ~ spray, data = InsectSprays, col = "lightgray",medlwd=1)
On Fri, Aug 22, 2008 at 3:18 PM, Suyan Tian <[EMAIL PROTECTED]> wrote:
> Hi, I just made a boxplot but I want to change the thickness of the line
> used
> -Original Message-
> From: [EMAIL PROTECTED]
> [mailto:[EMAIL PROTECTED] On Behalf Of Jay an
> Sent: Thursday, August 21, 2008 8:08 PM
> To: r-help@r-project.org
> Subject: [R] call perl
>
> Hi,
>
> It may be the old question.
> can anyone tell me how to call perl in R?
> # call perl i
On Fri, 22 Aug 2008, Kate Rohrbaugh wrote:
Greetings,
Apologies for such a simple question, but I have been trying to figure
this out for a few days now (I'm quite new to R), have read manuals,
list posts, etc., and not finding precisely what I need. I am
accustomed to working in Stata, but
Hi Avram --
My understanding is that Google-like map / reduce achieves throughput
by coordinating distributed calculation with distributed data.
snow, Rmpi, nws, etc provide a way of distributing calculations, but
don't help with coordinating distributed calculation with distributed
data.
SQL (a
Hi, I just made a boxplot but I want to change the thickness of the
line used for the median to look a little thinner. Could anyone please
help me figuring out the R-code to do it?
Thanks, I really appreciate it.
Suyan
__
R-help@r-project.org mail
On Thu, 21 Aug 2008, Roland Rau wrote:
Hi
Avram Aelony wrote: (in part)
1. How do others handle situations of large data sets (gigabytes,
terabytes) for analysis in R ?
I usually try to store the data in an SQLite database and interface via
functions from the packages RSQLite (and DBI).
I am testing whether the sample variances are equal. When p-value < 0.05
(alpha), should accept null hypothesis (sample variances are equal) or reject
it ?
The two new examples with each having same sample variances also puzzle me. Why
are the p-values different ?
bartlett.test(rep(rnorm(5),
If I understand correctly the result you are trying to achieve, I think
what you may be looking for is this:
model1.coeff <- lm(dv ~ iv1 + iv2 + iv3, data =
merged.dataset[merged.dataset$model1 == 1, ])
This will give you a regression using only the data for which model1 == 1.
on 08/22/2008
Have a look at
? cut
Best wishes
Wolfgang
--
Wolfgang Huber EBI/EMBL Cambridge UK http://www.ebi.ac.uk/huber
22/08/2008 14:14 Dr. G. Lawyer scripsit
>Hi,
>I have a mapping M consisting of a vector of bin centers
Dear Mark,
I would include the repeated measure as the smallest stratum in the random
effects specification:
random=~1|sampleunit/year
Setting up user-defined variance structures should be possible using for
example:
weights=varPower(form=~habitat)
or also try out the available corStruct()
I'm guessing that the following (untested)
does what is wanted:
function(x) {
pos <- sample(length(x), 2, replace=FALSE)
x[pos] <- x[ pos[2:1] ]
x
}
Patrick Burns
[EMAIL PROTECTED]
+44 (0)20 8525 0696
http://www.burns-stat.com
(home of S Poetry and "A Guide for the Unwilling S User")
The problem with trying to prove that 2 distributions are equal is that there
is exactly one way in which they can be equal and an infinite number of ways
that they can be different (and its rather a large infinity). The traditional
use of equality as the Null hypothesis works, because we can a
Hello,
We are attempting to use nlme to fit a linear mixed model to explain bird
abundance as a function of habitat:
lme(abundance~habitat-1,data=data,method="ML",random=~1|sampleunit)
The data consist of repeated counts of birds in sample units across multiple
years, and we have two questions:
What are your hypotheses? Once you state what they are, interpretation should
be straightforward.
-Original Message-
From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Daren Tan
Sent: Friday, August 22, 2008 11:18 AM
To: [EMAIL PROTECTED]
Subject: [R] Test of Homogeneity of
There are several options.
You can use the sunflowerplot function.
You can plot solid points with a small alpha value (not all graphics devices
support this), so that just a few points show up pale, but a lot show up as
more opaque.
You can use the symbols function (or bubbleplot, or my.symbol
Hello,
We would like to perform a cross validation on linear mixed models (lme) and
wonder if anyone knows of something analogous to cv.glm for such models?
Thanks, Mark
[[alternative HTML version deleted]]
__
R-help@r-project.org mailing list
I am testing the homogeneity of variances via bartlett.test and fligner.test.
Using the following example, how should I interpret the p-value in order to
accept or reject the null hypothesis ?
set.seed(5)
x <- rnorm(20)
bartlett.test(x, rep(1:5, each=4))
Bartlett test of homogeneity o
On Fri, 22 Aug 2008, Daniel Folkinshteyn wrote:
on 08/22/2008 10:19 AM Martin Ballaschk said the following:
how do I read files that have two header fields less than they have
columns? The easiest solution would be to insert one or two additional
header fields, but I have a lot of files and th
If I want to test for a salesman effect on average sale, I model it as "Sale ~
Salesman". But what if I want to test for a salesman effect on total sales?
How do I model that?
Given a significant salesman effect, how do I follow up with the equivalent of
Tukey's HSD to determine who the (indist
Exactly what do you want this statement to do:
if(merged.dataset$model1 == 1) model1.coeff <- lm(dv ~ iv1 + iv2 +
iv3, data = merged.dataset)
The result in the "if" is a vector of logical values (assuming that
merged.dataset is longer than 1); that is the source of the error
message. Do you want
on 08/22/2008 10:19 AM Martin Ballaschk said the following:
how do I read files that have two header fields less than they have
columns? The easiest solution would be to insert one or two additional
header fields, but I have a lot of files and that would be quite a lot
of awful work.
Any idea
Try this. It will read the file and see if there is a difference and
add in the extra headers:
x <- " time/ms C550.KMSCyt_b559.KMSCyt_b563.KMS
Cyt_f_.KMS P515FR.KMS Scatt.KMS Zea2.KMSPC
P700
0 Point1 -599.500 0.000 0.000
0.
Daniela Garavaglia wrote:
Sorry, I have some troubles with the graph device.
How can I draw a function over a histogram?
Thank's so much.
Daniela,
What function?
Here is one example using density() and using dnorm()
x <- rnorm(1000,2,2)
hist(x,prob=TRUE)
lines(density(x,na.rm=TRUE),col="red
Greetings,
Apologies for such a simple question, but I have been trying to figure this out
for a few days now (I'm quite new to R), have read manuals, list posts, etc.,
and not finding precisely what I need. I am accustomed to working in Stata,
but I am currently working through the multileve
Greetings,
Apologies for such a simple question, but I have been trying to figure this out
for a few days now (I'm quite new to R), have read manuals, list posts, etc.,
and not finding precisely what I need. I am accustomed to working in Stata,
but I am currently working through the multileve
Hi,
I have a mapping M consisting of a vector of bin centers (say 0.01, 0.02,
0.03 ..., 1) with associated values (say "red", "orange", "yellow", ...),
stored as a 2-column matrix with column 1 the bin center and column 2 the
value.
I also have a vector D of data (say 0.9056, 0.4
Sorry, I have some troubles with the graph device.
How can I draw a function over a histogram?
Thank's so much.
[[alternative HTML version deleted]]
__
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE d
Hello again,
I m trying to use timereg package as you suggested (R2.7.1 on XP Pro).
here is my script based on the example from timereg for a fine & gray model in
which
relt = time to event, rels = status 0/1/2 2=competing, 1=event of interest,
0=censored
random = covariate I want to test
lib
Hi,
how do I read files that have two header fields less than they have
columns? The easiest solution would be to insert one or two additional
header fields, but I have a lot of files and that would be quite a lot
of awful work.
Any ideas on how to solve that problem?
###
R stuff:
>
> Hello Richie,
> I would like to do three (or k) swap steps in each step just 2 ID
> recursive swaping
> x <- 1:10
> swap <- function(x){
> a <- sample(x,2)
> x[x==a[1]] <- swap[2]
> x[x==a[2]] <- swap[1]
> return(x)
> }
> swap(swap(swap(x))) -> mix
I tried my best with a response be
imicola schreef:
Hi,
I read somewhere that when carrying out geostatistical analysis in R you
should not use latitude and longitude...can anyone expand on this a little
for me, and what would be the best coordinate system to use?
I have my data in a geographic coordinate system, WGS84, decimal
Thanks a lot! I'll give it a closer look :-)
Ciao,
Antje
Gabor Grothendieck schrieb:
Look at the example section of ?new.env
Also the proto package on CRAN defines proto objects (which are just
environments with
slightly different semantics) and in the Applications section of the home page
h
Look at the example section of ?new.env
Also the proto package on CRAN defines proto objects (which are just
environments with
slightly different semantics) and in the Applications section of the home page
has links to examples of using proto objects to creates GUIs with gWidgets:
http://r-proto.g
Is this what you want:
> swap <- function(z){
+ a <- sample(length(z), 2)
+ z[a] <- z[rev(a)]
+ z
+ }
> swap(1:10)
[1] 2 1 3 4 5 6 7 8 9 10
>
> swap(1:10)
[1] 5 2 3 4 1 6 7 8 9 10
> swap(1:10)
[1] 8 2 3 4 5 6 7 1 9 10
> swap(1:10)
[1] 2 1 3 4 5 6 7
Hello again,
I m trying to use timereg package as suggested (R2.7.1 on XP Pro).
here is my script based on the example from timereg for a fine & gray model in
which
relt = time to event, rels = status 0/1/2 2=competing, 1=event of interest,
0=censored
random = covariate I want to test
Â
libr
Hello Richie,
I would like to do three (or k) swap steps in each step just 2 ID recursive
swaping
x <- 1:10
swap <- function(x){
 a <- sample(x,2)
 x[x==a[1]] <- swap[2]
 x[x==a[2]] <- swap[1]
 return(x)
 }
 swap(swap(swap(x))) -> mix
Â
Is this possible?
Thanks you in advance!
Amor
Hello,
I got the hint to use nesting environments (for a GUI problem), but I'm not
sure whether I understand what he means (and I did not really succeed with
reading documentation - probably I looked for the wrong keywords?).
That's what he wrote:
> Another good way to control the scoping is
Here is the way that I usually do it using lapply:
myData <- lapply(list.files(pattern=".RWL$"), function(.file)
read.rel(.file, header=TRUE))
# combine into a single dataframe
myDF <- do.call(rbind, myData)
On Thu, Aug 21, 2008 at 10:42 PM, Alison Macalady <[EMAIL PROTECTED]> wrote:
> Hi,
> I'
imicola wrote:
Hi,
I read somewhere that when carrying out geostatistical analysis in R you
should not use latitude and longitude...can anyone expand on this a little
for me, and what would be the best coordinate system to use?
Not only in R. In most systems, the inter-point distances are ass
This will also remove the NAs from the output; you will have to
change it to also keep the NAs. Wasn't sure what you wanted to do
with them.
dat <- data.frame(fac = rep(c("a", "b"), each = 100),
value = c(rnorm(130), rep(NA, 70)),
other = rnorm(200))
# split the
Dear all<>
I am running several generalized mixed model using lmer. <>The models
typical look like this:
model2xx<-lmer(numbers~Treatment+b+(1|Genotype)+(1|Field)+(1|Genotype:Treatment),
family=quasipoisson)<>
All factors are categorical.
<>And the output looks like this:
Generalized linear
Dear R-colleagues,
I am quite a newbie to R fighting my stupidity to solve a probably quite
simple problem of generating artificial data with defined features.
I am conducting a study of inter-observer-agreement in
child-bronchoscopy. One of the most important measures is Kappa
according to Fleis
On Thu, 2008-08-21 at 15:47 -0300, Paulo Barata wrote:
> Dear R list members,
>
> How to produce barplots anchored to the x-axis (not floating
> above the x-axis) with a box around?
>
> With both following codes, the lower horizontal line of the
> box is below the y = 0 line:
>
> # first code
>
Dear all, I'm doing a simulation study for an MM1 queue and I would like to
know if there are bias correction methods in R for time series (estimations of
warm-up points or something else).
Hope you can help me.
__
You could follow the advice given when loading the library (see the code you
posted for details) and use the enhanced genetics packages to cross-validate
the results (ideally you should get the same answer).
The results not weird though.
Your working with SNPs and having a homozygote with a freq
Maybe
?system
Might be what you need here?
-Rory
-Original Message-
From: [EMAIL PROTECTED] [mailto:[EMAIL PROTECTED] On Behalf Of Jay an
Sent: 22 August 2008 03:08
To: r-help@r-project.org
Subject: [R] call perl
Hi,
It may be the old question.
can anyone tell me how to call perl in
Hi
Apart from database interfaces such as sqldf which Gabor has mentioned, there
are also packages specifically for handling large data: see the "ff" package,
for instance.
I am currently playing with parallelizing R computations via Hadoop. I haven't
looked at PIG yet though.
Rory
-Ori
Hi,
I read somewhere that when carrying out geostatistical analysis in R you
should not use latitude and longitude...can anyone expand on this a little
for me, and what would be the best coordinate system to use?
I have my data in a geographic coordinate system, WGS84, decimal
degreesis this
> I wonder if there is any swap function in R that does the following:
> x <- seq(1,10,12)
This just makes x equal to 1. I'm guessing you meant something like
x <- 1:10
> x1 <- swap(x)
> x1
> 1 8 3 4 5 6 7 2 10
It's not terribly clear what you want swap to do. You can reorder the
elements of
Hello,
Something i don't understand
in your question.
Is treatment a time-dependent covariate?
That is, do patients receive the treatment
at the beginning of the study or later?
cmprsk cannot handle time-dependent covariates.
But if treatment is a baseline covariate,
but has a time-varying effe
Hi Rich,
Richard M. Heiberger wrote:
>
> Dan,
>
> The real problem is the use of csv files. csv files don't handle missing
> values
> ("#VALUE" is most likely from Excel), dates, or other complications very
> well.
>
> Read your Excel file directly into
> R with one of the packages designed
Hello everybody,
Â
I wonder if there is any swap function in R that does the following:
x <- seq(1,10,12)
x1 <- swap(x)
x1
1 8 3 4 5 6 7 2 10
Thank you very much!
Â
Amor
__
Schutz gegen Massenmails.
[[alternative HTML version deleted]
Paul,
thanks a lot for your advice! I got that kuss et al. paper and started to
read, this is more than I had expected! Very cool paper, and an algorithm
implemented in R that seems to be even superior to psignifit in matlab. So
thanks again!
mario
On 22-Aug-08 08:19:15, Marco Chiapello wrote:
> Hi,
> the chisq value for .05 level of signif. and 3 DF is 7.81 (one-sided).
> Is there a way to calculated it in R (giving level of signif. and DF)?
>
> Marco
qchisq(0.95,df=3)
# [1] 7.814728
qchisq(0.05,df=3,lower.tail=FALSE)
# [1] 7.814728
T
Hi everyone. My problem is about multivariate analysis. Given a multivariate
data X, of dimension (n by p), n is the number of observations while p is the
number of variables. I wanted a sample of size p+1 that will have its product
of the eigen values of the covariance matrix to be the minimum
1 - 100 of 113 matches
Mail list logo