I'm not as worried about the conn_to_db.print line, that is something I
have working ok, so if it looks funny please ignore it.
What I have so far, and its untested as of yet as I am trying to get the
form right, until I am sure its working .
Example:
while running
sleep 10
# read file into an array
f = file.open("mylogfile.log")
saved = File.open("archive.log", "a")
results = f.readlines
# Is there a way to remove everything that's been read here?
f.close
results.each_line do |sending|
conn_to_db.print("data #{sending}")
# Make sure we keep a backup
saved.puts("#{sending}") s.flush
end
# Close the archive file
saved.close
end
Thanks for any help!
--
Posted via http://www.ruby-forum.com/.
There's at least a couple of ways to remove duplicate lines.
1) results.uniq!
2) results.each { |l| b<<l } # b.keys gives you the set of unique lines
Hope that helps,
Jayanth
Sorry, I misread your question.
One silly way to achieve what you want is to reopen the file for reading.
That would pretty much erase the file. Not sure if that's what you want...
f.close
f = File.open("mylogfile.log","w")
f.close
Though, I don't see why you can't maintain a hash of all the lines instead
of having to remove lines from the log file. Its never a good idea to muck
about with log files.
Jayanth
I wanted to avoid that as I am using this log file to maintain results
from other scripts and if the timing goes wrong I might close the file,
another opens, then I open it again to erase and lose the latest entry.
Mostly I was wondering if there was a way to do this in one operation.
> Though, I don't see why you can't maintain a hash of all the lines
> instead
> of having to remove lines from the log file. Its never a good idea to
> muck
> about with log files.
hmmm...I suppose I could read the file and write out ones I recorded
with a notation so I skip those in the future. I'll have to think about
that.
Thanks