More on Disk Images -> Disk
Warren Toomey
wkt at henry.cs.adfa.oz.au
Wed Mar 25 10:32:56 AEST 1998
In article by Steven M. Schultz:
> Warren -
>
> >From: Warren Toomey <wkt at henry.cs.adfa.oz.au>
>
> > Now, what I've currently got will cope with -b12 compressed files. Can
> > someone tell me if it would be feasible to fit a gunzip into 64K?? Even
> > if it could only cope with gzip -1 files.
>
> If my understanding of 'gzip' is right then the alogrithm works on
> 32kb blocks of data and the '-N' level has little to do with the
> memory consumption. Rather, as the -1, ... -9 level increases the
> amount of work that gzip puts into the compression increases (the
> difference between -6 and -9 is only a few percent in final output
> size but the length of time taken is quite a bit higher).
>
> Of concern would be getting the gzip sources to compile with a non-ANSI
> compiler on a non-32bit machine (sizeof (long) == sizeof(int) is an
> endemic assumption I wager). Well, ok - there is the worry that
> you will grow old waiting for it to compress something ;-) Gzip is a
> lot more cpu intensive than compress.
I'm only thinking of implementing gunzip on the PDP-11. I've got
uncompress -b12 running standalone right now, but gunzip would be a big
win: you gzip -9 on a 32-bit system (higher compression) and gunzip
on the PDP-11.
I just don't know if the gunzip would fit. Isn't there a gunzip for MS-DOS?
Surely we could leverage something from it?
Warren
Received: (from major at localhost)
by minnie.cs.adfa.oz.au (8.8.5/8.8.5) id OAA20196
for pups-liszt; Wed, 25 Mar 1998 14:36:27 +1100 (EST)
X-Authentication-Warning: minnie.cs.adfa.oz.au: major set sender to owner-pups at minnie.cs.adfa.oz.au using -f
More information about the TUHS
mailing list