On Oct 15, 2009, at 3:43 AM, Biedermann, Jürgen wrote:
Hi there,
I'm facing the decision if it would be possible to transform several
more or less complex pdf files into an R Table-Format or if it has
to be done manually. I think it would be a impudent to expect a
complete solution, but I would be grateful if anyone could give me
an advice on how the structure of such a R-program could look like,
and if it's possible in general.
Here the problem:
Each pdf file belongs to a person. The pdf files actually represent
the anonymous criminal record of a person. Each entry should lead to
one row with the person number as key. The different lines should
form the columns. The criminal record actually looks like this:
---------------------------------------------------
Header with irrelevant text for us | Date: xx.xx.xxxx (relevant
for us)
Anonymous person number: xxxxxxxxxxx
Entries in the register
1. xx.xx.1902 -City-
Be in force since: xx.xx.1902
Date of offense:xx.xx.xxxx
Elements of the offence: For example "Rape"
Section in law: §176, §178 Abs. 1
Sentenced to 5 years imprisonment
"Irrelevant text for us"
Accommodation in an forensic psychiatry
Accommodation sentenced on probation
Rest of sentence sentenced on probation until the xx.xx.xxxx
2. xx.xx.1910
Be in force since: ....
.....
-----------------------------------------------------------------------
The problem is that the entries do not always have the same
structure. The first 6 lines are structurally the same in each entry
of the criminal record (each entry has a line for the judgement
date, the "be in force" date, the date of offence, the elements of
the offence, the Sections in law, and the sentence).
But then depending on the sentence different lines emerge which
contain information if the person was sentenced on probation, if the
probation was withdrawn again, when the person was released etc.
So, I think, these lines should be allocated to different columns
depending on key words. The definition of the key words for most
cases would not be the problem, actually. If a certain column is not
relevant in an entry (so, the key word didn't emerge) NA should be
put in the place.
But because sometimes (in rare cases), the entries contain spelling
errors, at the end, all the lines of an entry, which could not be
allocated to a column should be put in a column to check them
manually.
In the end the table should look more of less like this.
--------------------------------------------------
"Per
.Numb";"EntryNumber";"Judg.Date";"DateOffen.";...;"Probation.until";
"Released";"Not allocated"
xxxx1 1 xx.xx.1902 xx.xx.1901 ... xx.xx.1905 NA "blablabla"
xxxx1 2 xx.xx.1910 xx.xx.1909 ... NA 1925 "blablabla"
xxxx2 1 xx.xx.1924 xx.xx.1923 ... NA NA "blablabla"
------------------------------------------------------------------
Could anyone help me?
Thanks
Greetings
Jürgen
You don't indicate the OS you are on, but you will want to get a hold
of 'pdftotext', which is a command line application that can extract
the textual content from the PDF files. On most Linuxen, it is already
installed, but for Windows and OSX you will likely need to Google for
it.
The basic approach is to loop over each PDF file, use pdftotext to get
the text content and dump it into a regular text file. That file can
then be read into R using ?readLines.
This can all be done within R using the ?system command. Get the names
of the PDF files in a given folder by using ?list.files with a "\
\.pdf" or "\\.PDF" search pattern. Then ?paste together the full
command using a prefix along the lines of "pdftotext -layout -
nopgbrk", presuming that the pdftotext command is in your $PATH. The
suffix to be paste()d will be the name of the input PDF file and the
name of the output text file. So you end up with a command line
character vector along the lines of:
"pdftotext -layout -nopgbrk xxxxx.pdf xxxxx.txt"
where the x's are the specific file basenames. Review the pdftotext
options to understand what is being done and if you should need to
modify them for your particular files.
Once you have the data in R for each file, you will then need to
process the content line by line, looking for the keywords that are
associated with the content you require. Using ?grep is perhaps the
easiest way to accomplish that. You can then use ?gsub to replace/
strip the keywords, leaving you with the data only, for each line. For
multi line scenarios, you will need to keep track of where the keyword
for the first line is and then look for the subsequent keyword or
perhaps a blank line, to know when to stop aggregating the data for
that initial keyword.
It then becomes a matter of reorganizing the content that you need
into the format you require for subsequent work.
I have not looked for 'text processing' related packages on CRAN, so
you may wish to look there first in case there is anything relevant.
HTH,
Marc Schwartz
______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.