Google Groups no longer supports new Usenet posts or subscriptions. Historical content remains viewable.
Dismiss

split large file by string/regex

12 views
Skip to first unread message

Martin Dieringer

unread,
Nov 22, 2004, 3:38:55 AM11/22/04
to

I am trying to split a file by a fixed string.
The file is too large to just read it into a string and split this.
I could probably use a lexer but there maybe anything more simple?
thanks
m.

Steve Holden

unread,
Nov 22, 2004, 8:53:02 AM11/22/04
to
Martin Dieringer wrote:

Depends on your definition of "simple", I suppose. The problem with
*not* using a lexer is that you'd have to examine the file in a sequence
of overlapping chunks to make sure that a regex could pick up all
matches. For me that would be more complex than using a lexer, given the
excellent range of modules such as SPARK and PLY, to mention but two.

regards
Steve
--
http://www.holdenweb.com
http://pydish.holdenweb.com
Holden Web LLC +1 800 494 3119

Jason Rennie

unread,
Nov 22, 2004, 9:05:45 AM11/22/04
to pytho...@python.org

If the pattern is contained within a single line, do something like this:

import re
myre = re.compile(r'foo')
fh = open(f)
fh1 = open(f1,'w')
s = fh.readline()
while not myre.search(s):
fh1.write(s)
s = fh.readline()
fh1.close()
fh2.open(f1,'w')
while fh
fh2.write(s)
s = fh.readline()
fh2.close()
fh.close()

I'm doing this off the top of my head, so this code almost certainly
has bugs. Hopefully its enough to get you started... Note that only
one line is held in memory at any point in time. Oh, if there's a
chance that the pattern does not appear in the file, you'll need to
check for eof in the first while loop.

Jason

Diez B. Roggisch

unread,
Nov 22, 2004, 9:09:30 AM11/22/04
to
> Depends on your definition of "simple", I suppose. The problem with
> *not* using a lexer is that you'd have to examine the file in a sequence
> of overlapping chunks to make sure that a regex could pick up all
> matches. For me that would be more complex than using a lexer, given the
> excellent range of modules such as SPARK and PLY, to mention but two.

At least spark operates on whole strings if used as lexer/tokenizer - you
can of course feed it a lazy sequence of tokens by using a generator - but
that's up to you.

--
Regards,

Diez B. Roggisch

Martin Dieringer

unread,
Nov 22, 2004, 9:12:38 AM11/22/04
to
Steve Holden <st...@holdenweb.com> writes:

> Martin Dieringer wrote:
>
>> I am trying to split a file by a fixed string.
>> The file is too large to just read it into a string and split this.
>> I could probably use a lexer but there maybe anything more simple?
>> thanks
>> m.
>
> Depends on your definition of "simple", I suppose. The problem with
> *not* using a lexer is that you'd have to examine the file in a
> sequence of overlapping chunks to make sure that a regex could pick up
> all matches. For me that would be more complex than using a lexer,
> given the excellent range of modules such as SPARK and PLY, to mention
> but two.
>

yes lexing would be the simplest, but PLY also can't read from streams
and it looks to me (from the examples) as if it's the same with SPARK.
I wonder why something like this is not in any lib.
Is there any known lexer that can do this?
I don't have to parse, just write the junks to separate files.
I really hate doing that sequence thing...

m.

Martin Dieringer

unread,
Nov 22, 2004, 9:28:54 AM11/22/04
to
Jason Rennie <jre...@csail.mit.edu> writes:

> On Mon, Nov 22, 2004 at 09:38:55AM +0100, Martin Dieringer wrote:
>> I am trying to split a file by a fixed string.
>> The file is too large to just read it into a string and split this.
>> I could probably use a lexer but there maybe anything more simple?
>
> If the pattern is contained within a single line, do something like this:

Hmm it's binary data, I can't tell how long lines would be. OTOH a
line would certainly contain the pattern as it has no \n in it... and
the lines probably wouldn't be too large for memory...

m.

Bengt Richter

unread,
Nov 22, 2004, 12:21:15 PM11/22/04
to

Do you want to keep the splitting string? I.e., if you split with xxx
from '1231xxx45646xxx45646xxx78' do you want the long-file equivalent of

>>> '1231xxx45646xxx45646xxx78'.split('xxx')
['1231', '45646', '45646', '78']

or (I chose this for below)
['1231', 'xxx', '45646', 'xxx', '45646', 'xxx', '78']

or maybe

['1231xxx', '45646xxx', '45646xxx', '78']

??

Anyway, I'd use a generator to iterate through the file and look for the delimiter.
This is case-sensitive, BTW (practically untested ;-):

--< splitfile.py >----------------------------------------------
def splitfile(path, splitstr, chunksize=1024*64): # try a megabyte?
splen = len(splitstr)
chunks = iter(lambda f=open(path,'rb'):f.read(chunksize), '')
buf = ''
for chunk in chunks:
buf += chunk
start = end = 0
while end>=0 and len(buf)>=splen:
start, end = end, buf.find(splitstr, end)
if end>=0:
yield buf[start:end] #not including splitstr
yield splitstr # == buf[end:end+splen] # splitstr
end += splen
else:
buf = buf[start:]
break

yield buf

def test(*args):
for chunk in splitfile(*args):
print repr(chunk)

if __name__ == '__main__':
import sys
args = sys.argv[1:]
try:
if len(args)==3: args[2]=int(args[2])
except Exception:
raise SystemExit, 'Usage: python splitfile.py path splitstr [chunksize=64k]'
test(*args)
----------------------------------------------------------------

Extent of testing follows :-)

>>> print '%s\n%s%s'%('-'*40, open('splitfile.txt','rb').read(),'-'*40)
----------------------------------------
01234abc5678abc901234
567ab890abc
----------------------------------------
>>> import ut.splitfile
>>> ut.splitfile.test('splitfile.txt', 'abc')
'01234'
'abc'
'5678'
'abc'
'901234\r\n567ab890'
'abc'
'\r\n'
>>> ut.splitfile.test('splitfile.txt', '012')
''
'012'
'34abc5678abc9'
'012'
'34\r\n567ab890abc\r\n'
>>> it = ut.splitfile.splitfile('splitfile.txt','ab89',4)
>>> it.next
<method-wrapper object at 0x02EF1C6C>
>>> it.next()
'01234abc5678abc901234\r\n567'
>>> it.next()
'ab89'
>>> it.next()
'0abc\r\n'
>>> it.next()
Traceback (most recent call last):
File "<stdin>", line 1, in ?
StopIteration

(I put it in my ut package directory but you can put splitfile.py anywhere handy
and mod it to do what you need).

Regards,
Bengt Richter

Denis S. Otkidach

unread,
Nov 22, 2004, 1:20:11 PM11/22/04
to pytho...@python.org
On Mon, 22 Nov 2004 08:53:02 -0500
Steve Holden <st...@holdenweb.com> wrote:

> > I am trying to split a file by a fixed string.
> > The file is too large to just read it into a string and split this.
> > I could probably use a lexer but there maybe anything more simple?
> > thanks
> > m.
>
> Depends on your definition of "simple", I suppose. The problem with
> *not* using a lexer is that you'd have to examine the file in a sequence
> of overlapping chunks to make sure that a regex could pick up all

re module works fine with mmap-ed file, so no need to read it into memory.

> matches. For me that would be more complex than using a lexer, given the
> excellent range of modules such as SPARK and PLY, to mention but two.

--
Denis S. Otkidach
http://www.python.ru/ [ru]

William Park

unread,
Nov 22, 2004, 2:51:23 PM11/22/04
to

man strings (-o option)

--
William Park <openge...@yahoo.ca>
Linux solution for data management and processing.

Martin Dieringer

unread,
Nov 22, 2004, 2:48:16 PM11/22/04
to
"Denis S. Otkidach" <o...@strana.ru> writes:

> On Mon, 22 Nov 2004 08:53:02 -0500
> Steve Holden <st...@holdenweb.com> wrote:
>
>> > I am trying to split a file by a fixed string.
>> > The file is too large to just read it into a string and split this.
>> > I could probably use a lexer but there maybe anything more simple?
>> > thanks
>> > m.
>>
>> Depends on your definition of "simple", I suppose. The problem with
>> *not* using a lexer is that you'd have to examine the file in a sequence
>> of overlapping chunks to make sure that a regex could pick up all
>
> re module works fine with mmap-ed file, so no need to read it into memory.
>

thank you, this is the solution!
Now I can mmap.find all locations and then read the chunks them via
file.seek and file.read

m.

Martin Dieringer

unread,
Nov 22, 2004, 3:54:34 PM11/22/04
to
William Park <openge...@yahoo.ca> writes:

> Martin Dieringer <dier...@zedat.fu-berlin.de> wrote:
>> Jason Rennie <jre...@csail.mit.edu> writes:
>>
>> > On Mon, Nov 22, 2004 at 09:38:55AM +0100, Martin Dieringer wrote:
>> >> I am trying to split a file by a fixed string.
>> >> The file is too large to just read it into a string and split this.
>> >> I could probably use a lexer but there maybe anything more simple?
>> >
>> > If the pattern is contained within a single line, do something like this:
>>
>> Hmm it's binary data, I can't tell how long lines would be. OTOH a
>> line would certainly contain the pattern as it has no \n in it... and
>> the lines probably wouldn't be too large for memory...
>
> man strings (-o option)


this doesn't make sense at all

m.

Denis S. Otkidach

unread,
Nov 23, 2004, 6:00:51 AM11/23/04
to pytho...@python.org
On Mon, 22 Nov 2004 20:48:16 +0100
Martin Dieringer <dier...@zedat.fu-berlin.de> wrote:

> "Denis S. Otkidach" <o...@strana.ru> writes:

[...]


> > re module works fine with mmap-ed file, so no need to read it into
> > memory.
> >
>
> thank you, this is the solution!
> Now I can mmap.find all locations and then read the chunks them via
> file.seek and file.read

mmap-ed files also support subscription and slicing. I guess
mmfile[start:stop] would more readable.

Martin Dieringer

unread,
Nov 23, 2004, 11:22:23 AM11/23/04
to
"Denis S. Otkidach" <o...@strana.ru> writes:

> On Mon, 22 Nov 2004 20:48:16 +0100
> Martin Dieringer <dier...@zedat.fu-berlin.de> wrote:
>
>> "Denis S. Otkidach" <o...@strana.ru> writes:
> [...]
>> > re module works fine with mmap-ed file, so no need to read it into
>> > memory.
>> >
>>
>> thank you, this is the solution!
>> Now I can mmap.find all locations and then read the chunks them via
>> file.seek and file.read
>
> mmap-ed files also support subscription and slicing. I guess
> mmfile[start:stop] would more readable.

yes, even better :-)

m.

0 new messages