"Ignoramus15251" wrote in
I know, it's been confirmed to open properly. I was too sleepy to
check the modem error count. They recently switched me to a different
copper pair and cut it beyond my house, which reduced the line noise
considerably, but the wires are all the same age.
Do you have a file size of 91,265,098?
I finally went back to it and started the download, then told FF to
save the file as a PDF, then stopped the download to the browser. It
saved just fine and is working now. It just wouldn't read in the
browser for some reason.
[ ... ]
BTW -- Any chance of putting up "TDDR-50 F-4D trouble shooting manual
(1967)" (Navy, I think) if you know where to find it? I've been
going nuts trying to find that one for a project I'm working on.
(It is a manual for the F-4D fighter aircraft, and I'm looking
for it to find the pinout information on an HSI (Horizontal
Situation Indicator). I've buzzed out quite a bit of
information about the 96 leads (two connectors -- 41-pin and
55-pin) and some missing amplifiers which the smaller connector
should go to.
[ ... ]
Hmm ... wget is predicting about 8 minutes on a T1 connection.
I'll go play a game while that is downloading to see whether what I get
Download time was: 8m 23s
O.K. 292 pages, with all that I (randomly) checked readable, so
the file is not bad at all.
First try download was 8m 23s with "wget". Displays fine with
"xpdf". Same with Acrobat reader. *But* you need to get the whole file
before it will work. I think that PDF puts some critical information at
the end of the file, so if you don't get it all, you can't view it. 87
MB is the short form, the actual byte count was 91,265,098.
I think that you need to re-try -- and maybe you'll get a better
routing this time.
I did. Earthlink dropped out twice last night after downloading
20-30MB. This morning I lost them while browsing, but BasicISP is
fine. However their 4-hour timeout limits downloads to about 65MB.
Neither of them is completely reliable here, nor is the cellular
I'm using Linux now but the Windows binaries always worked fine for me.
This will let you download partial files and automatically restart
downloads when the server drops a connection. I tested Iggy's
server/file and it does allow you to stop/restart downloads. Not all
servers are set up that way though.
Try taking the partial download you currently have and while having a
Windows command prompt open in that directory type:
=== from wgets manual
Continue getting a partially-downloaded file. This is
useful when you want to finish up a download started by a previous
instance of Wget, or by another program. For instance:
wget -c ftp://sunsite.doc.ic.ac.uk/ls-lR.Z
If there is a file named ls-lR.Z in the current directory,
Wget will assume that it is the first portion of the remote file, and
will ask the server to continue the retrieval from an offset equal to
the length of the local file.
Note that you don't need to specify this option if you just
want the current invocation of Wget to retry downloading a file should
the connection be lost midway through. This is the default behavior.
-c only affects resumption of downloads started prior to this
invocation of Wget, and whose local files are still sitting around.
Without -c, the previous example would just download the
remote file to ls-lR.Z.1, leaving the truncated ls-lR.Z file alone...
Trust me, give it a whirl ;-)
Too big for whom? It depends on what program you use to view
it, and how much memory your computer has. So that warning would be
useless for some, and likely ignored by others. IIRC, the download site
tells you the size. It is then up to *you* to know whether it is too
big, with *your* computer's memory and such, to view the file.
Today's browsers try to view in a java program in the browser,
instead of downloading it and passing it to a stand-alone program which
only views PDF files.
You can configure most browsers to download and save instead of
attempt to display it -- and you can download a copy of Adobe Acrobat
for most systems to view the file. Download it, save it, and then exit
the memory hog browser and start up Adobe Acrobat (or whatever other PDF
reader you have which is *not* part of a browser.
I need a new computer. I'm up to 2G of RAM, but that's not enough for
Win7 and all the windows I have open on a daily basis. I was running
97% full memory this evening when opening extra tabs in FF for all the
goodies in today's Gizmag email.
In the past, Windows has created ramdisk memory to store files larger
than viewable with existing RAM.
I believe I have Acrobat Reader for Firefox, which should have worked,
but it didn't. Must check into that...
I have Acrobat Reader, which successfully opened it after download.
I'll check it out. AR has been giving me "may be farked" messages on
3/4 of the downloads lately, but they have been displaying flawlessly.
P.S: See how easy it is to snip, and include only the relevant info?
It took maybe 3 seconds to do.
I ran XP without any updates from March 2013 to March 2014 and had no
problems. AVG caught a few things it didn't like.
I make Seagate Disk Wizard backups of C: to DVD-DLs that can be booted
and restored from a read-only DVD drive. A virus would have to hide in
the BIOS or a peripheral to get back in after that, since I don't
network my on-line PC to the others, and the strict restrictions on
the account I browse with make that difficult.
Gunner Asch on Tue, 17 Jun 2014 07:36:43 -0700
typed in rec.crafts.metalworking the following:
That was MS position when they were supporting XP, especially if
you did not use their anti-virus program.
"With Age comes Wisdom. Although more often, Age travels alone."