Jeff Abrahamson on 17 Oct 2006 23:19:12 -0000


[Date Prev] [Date Next] [Thread Prev] [Thread Next] [Date Index] [Thread Index]

Re: [PLUG] how to transfer gigabytes without tech support


Yeah, that's it.  I was having him use a machine that only allows
certificates.  I just needed to switch to a machine that allows
passwords.  Same file system.  I just created an account for him and
made a symlink to where the files are, then gave him an rsync command
to copy and paste to terminal.

*Whew*.  I knew there had to be a way around this foolishness.

Thanks.

-Jeff


On Tue, Oct 17, 2006 at 07:01:49PM -0400, Chad Vogelsong wrote:
>   [55 lines, 320 words, 2333 characters]  Top characters: _etoasn-
> 
> Option #1:
> 
> Setup and sFTP server on your machine (I use vsftpd).  He should only need to 
> type a command, a password and another command to download it.  Just hope that 
> there are no cuts in service during the days it will take to transfer that 
> amount of data over a Cable/DSL connection (slow upload speed usually).
> 
> He would just have to type a command like:
> $ ftp username@server.ip.address
> password:  the-password-you-give-him
> get filename.tar.gz
> 
> Wait 1-3 days, depending on speed.
> 
> Option #2
> 
> Another way would be to give him an SSH account on your computer.  Put the file 
> in his home directory.  Then it's just 1 command he has to type.
> 
> scp username@server.ip.address:~/filename.tar.gz ~/
> password: the-password-you-give-him
> 
> Since you probably already have SSH installed on your computer, the 2nd one is 
> probably easier.  It's also more secure IMHO.
> 
> -Chad
> 
> 
> Jeff Abrahamson wrote:
> > I have 14 GB of data I want to give someone.  He's on MacOS X.  I
> > thought to do rsync via ssh, but talking him through certificates and
> > terminal things is becoming too painful.  Simple things fail that I
> > would see if I were there, but I'm not and his descriptions omit
> > little details that he has no way of knowing are critical.
> > 
> > Next thought, make it available for http.  So I tar it up.  But apache
> > refuses to serve a 14 GB tar archive, saying it's too big.  (Weirdly,
> > this causes a directory to disappear, too: directory foo/ is visible
> > until overly large foo.tar is present, then they both disappear from
> > the directory listing that apache generates!)
> > 
> > Any thoughts on how to do this as simply as possible?
> > 
> > 
> > 
> > ------------------------------------------------------------------------
> > 
> > ___________________________________________________________________________
> > Philadelphia Linux Users Group         --        http://www.phillylinux.org
> > Announcements - http://lists.phillylinux.org/mailman/listinfo/plug-announce
> > General Discussion  --   http://lists.phillylinux.org/mailman/listinfo/plug
> ___________________________________________________________________________
> Philadelphia Linux Users Group         --        http://www.phillylinux.org
> Announcements - http://lists.phillylinux.org/mailman/listinfo/plug-announce
> General Discussion  --   http://lists.phillylinux.org/mailman/listinfo/plug

-- 
 Jeff

 Jeff Abrahamson  <http://jeff.purple.com/>          +1 215/837-2287
 GPG fingerprint: 1A1A BA95 D082 A558 A276  63C6 16BF 8C4C 0D1D AE4B

Attachment: signature.asc
Description: Digital signature

___________________________________________________________________________
Philadelphia Linux Users Group         --        http://www.phillylinux.org
Announcements - http://lists.phillylinux.org/mailman/listinfo/plug-announce
General Discussion  --   http://lists.phillylinux.org/mailman/listinfo/plug