[news.admin] some

werner@utastro.UUCP (Werner Uhrig) (05/22/88)

	[ this article was crossposted to news.admin because I feel that the ]
	[ basic premise applies to non-Macintosh groups also ]
Folks,
	I, usually, pipe hexed articles through xbin to take advantage of the
reduced download-time of the resulting unhexed file(s).  

I was rather bothered that xbin (on my favorite SUN) choked on GuardDog,
enough so, that I decided to figure out why.  I downloaded the hex-sources
and unhexed them with BinHex4, resulting in a file called "Guard Dog TM"
(where TM represents the sign for Trademark).  Suspecting that the non-ASCII
character may be the cause of xbin choking, I changed the name to "Guard Dog"
(given that 'Guard Dog' is of type SYSTEM you have to 'bend some bits'
before you can rename the file), hexed and uploaded that file, and, presto,
xbin stopped dumping core.

While at it, I figured I might as well check how much smaller compressed file 
would have been (I used StuffIt-1.40 for that), with the following
result:

  32 -rw-r--r--  1 werner   other       32134 May 21 17:36 Guard_Dog.Hqx
  18 -rw-r--r--  1 werner   other       18336 May 21 17:55 Guard_Dog.sit.hqx

All this leads me to the following suggestion:

ALL files distributed on COMP.BINARIES.MAC and COMP.SOURCES.MAC should be
COMPRESSED before being HEXED.  The (intermediate) compressed file should
have a name consisting of ASCII-characters only so that xbin can be used to
unhexify the article on the mainframe in order to reduce the download time.

It would be best if the person submitting the article would do "the work",
however, the moderator should, at least, verify that the submitted article
conform to the standard  (I assume that he already verifies that no pirated
material gets distributed, by inspecting the submitted material).

If the moderator is unable (or unwilling) for technical or personal reasons
to take on that (additional?) work, I propose that a call go out for
volunteers to help with such tasks - or for a new moderator who is able and
willing to take on or coordinate such an effort.

		Cheers,		---Werner

	"Let's take the best there is and improve it"
		--- someone famous is bound to have come to that conclusion
			before me ....

PS: if someone has a version of xbin that can handle ALL non-ASCII chars in
	file-names, I'd appreciate .....

macak@lakesys.UUCP (Jim Macak) (05/24/88)

In article <2689@utastro.UUCP> werner@utastro.UUCP (Werner Uhrig) writes:
> (much deleted...)
>
>ALL files distributed on COMP.BINARIES.MAC and COMP.SOURCES.MAC should be
>COMPRESSED before being HEXED.  The (intermediate) compressed file should
>have a name consisting of ASCII-characters only so that xbin can be used to
>unhexify the article on the mainframe in order to reduce the download time.
>
> (etc.)

Yes, an excellent suggestion.  It's crazy to have 10 part files posted in
comp.binaries.mac when file compression can reduce those down to 50-60% of
that size.

And while we're at it, perhaps a standard program for doing the
packing/compression ought to be adopted.  Ray Lau's StuffIt is the obvious
choice.  Though still with an aoccasional bug, StuffIt is certainly stable
enough to use for mere stuffing/unstuffing of files.  StuffIt has certainly
become the standard on the GEnie, CIS and other Mac sections, so why not here
too?  Besides, Ray Lau only requests a shareware fee if one uses StuffIt for
encoding/archiving purposes.  Otherwise, it is considered a freebie: all the
more reason to adopt it as a standard on this system.

Jim

-- 

>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>><<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<
Jim -->  macak@lakesys.UUCP (Jim Macak)  {Standard disclaimer, nothin' fancy!}
>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>>><<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<<

gz@spt.entity.com (Gail Zacharias) (05/24/88)

In article <699@lakesys.UUCP> macak@lakesys.UUCP (Jim Macak) writes:
>And while we're at it, perhaps a standard program for doing the
>packing/compression ought to be adopted.  Ray Lau's StuffIt is the obvious
>choice. <....> Besides, Ray Lau only requests a shareware fee if one uses
>StuffIt for encoding/archiving purposes.  Otherwise, it is considered a
>freebie: all the more reason to adopt it as a standard on this system.

In order for you to use this "freebie", the contributor has to use StuffIt
for encoding.  This pretty much excludes people like me, who oppose shareware
on principle, from contributing.

I suggest tar and compress instead.  There are (several) freeware versions of
each available for the macintosh, and I'm sure it'd be easy to write versions
for other micros.  And besides, every unix system has them, so you're not
locked into one individual's good will for continued support.  This also gives
archive sites (which are pretty much all Unix) a lot of flexibility on how to
handle storage without having to transfer files to a micro for repackaging.
For example, the software could be posted as uuencoded tar files, then
automatically uudecoded and compressed on receipt at the archival host,
thereby getting around the compressing-compressed-files problem of news
transmission.
 --
gz@entity.com					 ...!mit-eddie!gz
	   Unix: when you can't afford the very best.

bytebug@dhw68k.cts.com (Roger L. Long) (05/25/88)

In article <699@lakesys.UUCP> macak@lakesys.UUCP (Jim Macak) suggests
standardizing using StuffIt for packing/compression.

In article <307@spt.entity.com> gz@eddie.mit.edu (Gail Zacharias) objects:
>In order for you to use this "freebie", the contributor has to use StuffIt
>for encoding.  This pretty much excludes people like me, who oppose shareware
>on principle, from contributing.

Actually not.  That's the job of the moderator.  People could contribute
stuff in just about any reasonable form that I could figure out how to deal
with, and I'd repackage things into the "standard" form here.  An there is
a public domain (at least for UNIX) un-StuffIt package, so no reason to 
even USE the shareware package if you don't care to.

What matters most (to me) for submitting things to comp.{binaries,sources}.mac
is that 

  -  the submission gets here intact
  -  it is real obvious who the submitter is (i.e. name, email address,
     organization, all the stuff that makes up an article header).
  -  if there is a version number, it'd be nice if you'd include that in
     the header, since one complaint a number of people have is being
     aware of what version something is, so they don't have to download
     it to see if it's an updated version of something they already have.
  -  Documentation, if provided, is TEXT or MacWrite, since these seem
     to be most standard.  Besides, I don't personally have every
     Word Processing package available, so I can't check out documentation
     that comes in other flavors.
  -  You write a paragraph or so explaining what you are posting (yes, I
     take a look at what's posted, but have better things to do than
     write a description for something that YOU felt was worth posting).
     Remember, in this paragraph, you want to convey enough information
     to the people who receive the article to give them something by which
     they can decide if they want to spend their time downloading it.
     Let people know if documentation is included.  If something will only
     work on a Mac II, let people know that in the text description, so
     that people with Mac SE's don't waste their time downloading something
     that has no hope of working on their machine.  If it requires the
     latest version of the system software, let people know that so they
     will understand why things don't work on their machine if they're using
     old software.

Thanks.

-- 
	Roger L. Long
	dhw68k!bytebug

earleh@eleazar.dartmouth.edu (Earle R. Horton) (05/26/88)

In article <2689@utastro.UUCP> werner@utastro.UUCP (Werner Uhrig) writes:
>
> (He says xbin doesn't like file name characters > 0x7F)
>
>PS: if someone has a version of xbin that can handle ALL non-ASCII chars in
>	file-names, I'd appreciate .....

Werner requests that posters of Macintosh programs not use non-ASCII
characters in file names, since these are not legal in UNIX file names.
It is better to fix xbin, I think, since then you don't place artificial
restrictions on Macintosh users, who may not especially care about UNIX
peculiarities.  The fix is easy, just AND every character in the UNIX
file names used by xbin with 0x7F.  Here is a C code fragment from my
copy of xbin.  Note that I have added translation for some (but not all)
characters that sh and csh do not like.  This is, frankly, getting
cumbersome, and I am thinking of converting to use of a translation
table someday...

        namebuf[n] = '\0';
 
        /* get rid of troublesome characters */
        for (np = namebuf; *np; np++){
                if (*np == ' ' || *np == '/' || *np == '!' ||
                *np == '(' || *np == ')' || *np == '[' || *np == ']'
                || *np == '*' || *np == '<' || *np == '>' ||
                *np == '?' || *np == '\'' || *np == '"' || *np == '$')
                        *np = '_';
                *np &= 0x7f;
        }
        sprintf(files.f_data, "%s.data", namebuf);
        sprintf(files.f_rsrc, "%s.rsrc", namebuf);
        sprintf(files.f_info, "%s.info", namebuf);

My 2 cents worth:

I don't think StuffIt is a good choice for a standard of compressing
files, since then every poster has to shell out $20.00 to be a legal
user.  I am partial to use of UNIX-compatible tar and compress.  There
exist both tar and compress for the Macintosh, both freeware, but
only the MPW Tool version of tar handles the resource fork.  At this
time, there is not a suitable method of archiving/compressing
Macintosh files which:

	a)  Is free.
	b)  Runs under the Finder.

The MPW version of Tar can archive Macintosh files using the MacBinary
standard, and the resulting file can be unTar'ed on any UNIX machine
and possibly VMS and other systems, also on any Macintosh which has
MPW installed.  Portions of the archive can be downloaded to a
Macintosh from any system which has an xmodem program, and many free
terminal programs will recognize the MacBinary format of the
downloaded file(s).  Tar files can be compressed using MacCompress2.1,
which is compatible with UNIX compress.  If the MPW Tool version of
Tar were compiled into a Macintosh application, then I think this
would be an ideal choice for an all-purpose archiving program.
Volunteers?
*********************************************************************
*Earle R. Horton, H.B. 8000, Dartmouth College, Hanover, NH 03755   *
*********************************************************************

straka@ihlpf.ATT.COM (Straka) (05/27/88)

In article <8604@dartvax.Dartmouth.EDU> earleh@eleazar.dartmouth.edu (Earle R. Horton) writes:
>My 2 cents worth:
>
>I don't think StuffIt is a good choice for a standard of compressing
>files, since then every poster has to shell out $20.00 to be a legal
>user.  I am partial to use of UNIX-compatible tar and compress.  There

There are two issues here (as least as far as the net is comcerned.  One is
getting the binary to the moderator.  Compression is not crucial here.  Any
means of getting the binary to the moderator would not stress the bandwidth
of the network too much.

The moderator unpacks, unstuffits, or unhexifies the binaries, reviews them,
and only THEN stuffits them together for distribution on the net.  This way,
only the moderator is obliged to contribute the shareware fee.

I don't think there is much of an argument against Stuffit.  Of course, if
you want to help out the revenues of AT&T, Sprint, MCI, et al., then go right
ahead...  I would prefer to use the same resources to process MORE binaries
through the net for the same cost using Stuffit.
-- 
Rich Straka     ihnp4!ihlpf!straka

Advice for the day: "MSDOS - just say no."

mclek@dcatla.UUCP (Larry E. Kollar) (05/28/88)

[I'm cross-posting to comp.sys.amiga since this applies to amiga binaries
as well.]

To reduce the amount of binaries on the net, and to enhance the usefulness of
what's posted, I propose the following scoring system:

	All submissions have a base score of 0.

	Sources accompany submission:  +10
	Public domain (as opposed to free copyrighted): +5
	Binary < 20K:  +5
	Binary < 50K:  +3 (a 10K binary has a score of +5, not +8)

	Binary > 100K:  -3
	Binary > 150K:  -5
	Shareware:  -5
	Demo of commercial program:  -10

Each submission is given a score, and placed in the posting queue based on that
score.  The higher the score, the sooner it gets posted.  Small things aren't
clobbered nearly as often as larger programs, and programs with sources are
much more useful, so we should encourage these submissions.  Big shareware or
commercial demos would most likely never get posted, due to things jumping
ahead of them in line.  The Mac and Amiga are hard enough to learn to program;
the until-recent dearth of example Mac sources didn't help.

These, of course, are rough guidelines; since moderators are intelligent humans,
they can modify scores to their own tastes or for special considerations (if
everyone is asking for it, it should be posted to reduce request clutter).

On the other hand, people could port GNU CC to the Amiga & Mac, then we could
do away with binaries entirely. :-)

Well, what do y'all think?

	Larry Kollar	...!gatech!dcatla!mclek

sean@ms.uky.edu (Sean Casey) (05/29/88)

In article <5145@dcatla.UUCP> mclek@dcatla.UUCP (Larry E. Kollar) writes:
>To reduce the amount of binaries on the net, and to enhance the usefulness of
>what's posted, I propose the following scoring system:

I hope no one takes this seriously.

Sean
-- 
***  Sean Casey                        sean@ms.uky.edu,  sean@ukma.bitnet
***  The Empire Definistrator          {rutgers,uunet,cbosgd}!ukma!sean
***  ``I'm not gonna mail it, YOU mail it. I'M not gonna mail it... Hey! Let's
***  sent it to Rutgers! Yeah! They won't mail it! They return everything...''

jmpiazza@sunybcs.uucp (Joseph M. Piazza) (05/30/88)

In article <5145@dcatla.UUCP> 	Larry Kollar writes:
>To reduce the amount of binaries on the net,

	What ever for? ...

> ... and to enhance the usefulness of
>what's posted, I propose the following scoring system:
>
>	All submissions have a base score of 0.
>
>	Sources accompany submission:  +10

	... You seem to consider Sources intrinsically more valuable than
Binaries.  While I'm sure you have reasons, some true and good,  it can't
hold true for many situations.  For one thing, this subject has been thrashed
about many times in many news-groups and yet binarie news-groups still exist.

	The fact remains that sources are (usually) larger than binaries --
no savings here.

	Not everybody has the apropriate compiler/interpreter/assembler.
Some people may have them but may not know enough to make everything work.  A
good example is if the user doesn't own the right brand compiler.  This
sometimes includes me.  Do you know what that you would be saying to me?
"Tough shit."

>... programs with sources are
>much more useful, so we should encourage these submissions.

	I do agree that sources  c a n  be more useful ... (hold this thought)

>... Big shareware or
>commercial demos would most likely never get posted, due to things jumping
>ahead of them in line.

	There's smoething slippery about this idea.  An unpredictable delay
of a posting could deminish its usefulness.  This could also cause a
sufficietly delayed posting prove virtually useless and therefore a
detriment -- but not dependent on the posting's own merit.  There's no good
in that.

>... The Mac and Amiga are hard enough to learn to program;

	... And some people don't program at all.  (Remember that thought)
Should we ignore non-programmers?

>...
>These, of course, are rough guidelines; since moderators are intelligent
>humans, they can modify scores to their own tastes or for special
>considerations ...

	I think it we would best leave it the hands of intelligent humans
than to a non-thinking scoring scheme -- we're not playing Bridge.

Flip side,

	joe piazza

---
In capitalism, man exploits man.
In communism, it's the other way around.

CS Dept. SUNY at Buffalo 14260
UUCP: ..!{ames,boulder,decvax,rutgers}!sunybcs!jmpiazza         GEnie:jmpiazza
BITNET: jmpiazza@sunybcs.BITNET         Internet: jmpiazza@cs.Buffalo.edu

>	Larry Kollar	...!gatech!dcatla!mclek

greg@gryphon.CTS.COM (Greg Laskin) (05/30/88)

In article <11637@sunybcs.UUCP> jmpiazza@sunybcs.UUCP (Joseph M. Piazza) writes:
>In article <5145@dcatla.UUCP> 	Larry Kollar writes:
>>To reduce the amount of binaries on the net,
>	What ever for? ...
>
>	... You seem to consider Sources intrinsically more valuable than
>Binaries.  
>	Not everybody has the apropriate compiler/interpreter/assembler.

The question is really whether we're just going to grab everything off
every BBS and redistribute it over USENET.  <-- hyperbole -->

Source code is usually the original work of the poster as are most of
the other postings to USENET.  It's not difficult to build a case for
distributing a binary version of, say, uemacs for those without
the proper compiler because uemacs is, by and large, a USENET "product."
Therefore, it's not difficult to make a case for the binary groups.

Third party postings of binaries culled from other information providers
are more problematical.  Such usage may be outside the scope of the
"reasonable" amount of service that network sites might be expected to provide.
In fact, I feel much the same about some of the other high volume
retransmission by third parties that I've seen go by in the last
few months and not just binaries.

I don't recall much discussion of this point.  What do we think?
-- 
Greg Laskin  greg@gryphon.CTS.COM    <any backbone site>!gryphon!greg

chip@vector.UUCP (Chip Rosenthal) (05/30/88)

In article <11637@sunybcs.UUCP> jmpiazza@sunybcs.UUCP (Joseph M. Piazza) writes:
>In article <5145@dcatla.UUCP> 	Larry Kollar writes:
>>... The Mac and Amiga are hard enough to learn to program;
>	... And some people don't program at all.  (Remember that thought)
>Should we ignore non-programmers?

Yes.
-- 
Chip Rosenthal /// chip@vector.UUCP /// Dallas Semiconductor /// 214-450-0400
{uunet!warble,sun!texsun!rpp386,killer}!vector!chip
I won't sing for politicians.  Ain't singing for Spuds.  This note's for you.

dyker@boulder.Colorado.EDU (Barbara Dyker) (06/01/88)

In article <8297@dhw68k.cts.com> bytebug@dhw68k.cts.com (Roger L. Long) writes:
>In article <699@lakesys.UUCP> macak@lakesys.UUCP (Jim Macak) suggests
>standardizing using StuffIt for packing/compression.
>
>What matters most (to me) for submitting things to comp.{binaries,sources}.mac
>is that 
> [long list of good guidlines deleted]


What would also take some load off the net would be a REGULAR posting in
comp.binaries.* of a list of submittal guidlines AND the procedures AND
software required to unpack/hex the posting!!!

The net is all too often cluttered with requests for BinHex, StuffIt, PackIt...
and hasty articles about not being able to unbinhex 4.0 with 5.0, and "what's
a .pit file?"

I did my stuggling as a newbie - let's get some info out for those that
are new to the net so that it works for all of us.

Or shall we ignore newbies like someone suggested we ignore non-programmers??


Barb Dyker		CSNET:	dyker@boulder.Colorado.EDU
			UUNET:	...rutgers!ncar!dinl!tosgcla!dyker
"Do what you will with my employer, but leave me alone."

lmb@vsi1.UUCP (Larry Blair) (06/09/88)

Sorry to join this discussion at such a late date.  I had been ignoring it, but
while flipping through I saw the word "compress" and that disturbed me.

The vast majority of news is transported site to site compressed.  Applying
compression to postings will most likely result in an increase data transfered.

lmb@vsi1.UUCP (Larry Blair) (06/09/88)

In article <8845@dartvax.Dartmouth.EDU> earleh@eleazar.dartmouth.edu (Earle R. Horton) writes:
|In article <643@vsi1.UUCP> lmb@vsi1.UUCP (Larry Blair) writes:
|>The vast majority of news is transported site to site compressed.
|>Applying compression to postings will most likely result in an
|>increase data transfered.
|
|Example, please:
|
|      No StuffIt Used          |      Stuffit Used
|                               |
| Stage            Size (bytes) | Stage                 Size
| -----            -----------  | -----                 ----
|                               |
|Application           33756    | Application           33756
|Application.Hqx       45082    | Application.sit       24070
|Application.Hqx.Z     31395    | Application.sit.Hqx   32896
|                               | Application.sit.Hqx.Z 29683

I stand corrected.  These results seem to contradict the notion that
compression randomizes a file such that further compression is useless.
I guess that the ASCII-tizing of the compressed data adds enough non-
randomness to allow more compression.
-- 
*   *   O     Larry Blair
  *   *   O   VICOM Systems Inc.     sun!pyramid----\
    *   *   O 2520 Junction Ave.     uunet!ubvax----->!vsi1!lmb
  *   *   O   San Jose, CA  95134    ucbvax!tolerant/
*   *   O     +1-408-432-8660

werner@utastro.UUCP (Werner Uhrig) (06/10/88)

In article <643@vsi1.UUCP>, lmb@vsi1.UUCP (Larry Blair) writes:
> while flipping through I saw the word "compress" and that disturbed me.
 
> The vast majority of news is transported site to site compressed.  Applying
> compression to postings will most likely result in an increase data transfered

will this baloney-argument never stop?  are you telling me that when I reduce
the size of a text file (or whatever) by 40% by compressing it, here comes
the transmission program trying to compress it, only to double it back in size?

I'd be surprised if the worst case of negative pay-off of having integrated
compression into the transmission protocols are more than a few % at most;
otherwise, there must be something terribly wrong with it.  If it can't
be smart about supporting people who want to reduce disk-usage for all these
articles queueing up on thousands of USEnet machines (well, give it a little
help with an additional header "Compressed-Article: <algotrithm used>" if
you must) then I can't understand why we bother at all ....

If it could be shown that, overall, so many articles are now precompressed
that transmission-integrated compression no longer buys anything - then
we have reached the ideal state where compression on-the-fly during transmission
is no longer needed ... but I don't buy that "proof-unseen"!

if someone has studied actual data and can present ideas/suggestions based
on the analysis, I'm sure that's worth discussing. otherwise, ....

farren@gethen.UUCP (Michael J. Farren) (06/18/88)

In article <2758@utastro.UUCP> werner@utastro.UUCP (Werner Uhrig) writes:
>
>will this baloney-argument never stop?  are you telling me that when I reduce
>the size of a text file (or whatever) by 40% by compressing it, here comes
>the transmission program trying to compress it, only to double it back in size?

The compress program used in Usenet transmission will not result in a 
larger transmitted file than that stored.  But - what happens when your
pre-compressed file gets sent through a mailer which doesn't understand
eight-bit data?  To guarantee transmission throughout the entire net, 
you need to convert a compressed file into pure ASCII by using uuencode
or some such, and this DOES increase file size significantly.

Remember, too, that compression is quite ineffective on smaller files
(such as individual articles), and fairly CPU-intensive as
well.  The standard news transmission scheme batches many articles together
before compressing, resulting in much greater efficiency. 

-- 
Michael J. Farren             | "INVESTIGATE your point of view, don't just 
{ucbvax, uunet, hoptoad}!     | dogmatize it!  Reflect on it and re-evaluate
        unisoft!gethen!farren | it.  You may want to change your mind someday."
gethen!farren@lll-winken.llnl.gov ----- Tom Reingold, from alt.flame