[ARGUS] argus configuration here ...

Peter Van Epp vanepp at sfu.ca
Thu Oct 14 12:27:19 EDT 2004


	Mike asked off list if I'd share my current config, and since I expect
perhaps others may be interested I'll do it on list. We have 3 major campuses
around town (and another 3 or 4 more minor presences) pretty much fully 
interconnected with dark fibre. We are also one of the 5 owners of the local
regional optical network (who technically owns the dark fibre) which in turn
is currently the GigaPop operator for the CA*net4 presence in BC. The Gigapop
resides in space leased at our downtown campus. We have a clear channel Gig
link in to the Gigapop and thus CA*net4/I2 in the US and a 100 meg connection 
with a packeteer on it to control cost :-) in the to the transit exchange
at the same place for commodity traffic. As well we have a grid computing 
research project that houses a 200 terabyte file store and some smaller 
(~192 nodes beowulf cluster, an IBM G4 blade server, soon a Cray) here, and
major computing nodes (1500 node beowulf cluster, large SGI, large Alpha 
I think) at other Universities both here in town and in Alberta (one province
over). They interconnect on a current 1 gig light path via C4 (the C4 backbone
is OC192) on its way to 10 gigs at least locally. That link is currently 
"testing" one of my argus boxes to make sure my Syskonnect cards are OK (they 
had borrowed my machine and had trouble with Linux and the SysKonnect cards,
they obviously need to upgrade to a real operating system :-)) and should be 
the most interesting because it has been known to peak at ~950 megabits per 
second (although it hasn't done so since argus has been on). As well one of 
our remote campuses doesn't yet have fibre and so is limping along on a 10 meg 
FDX circuit and have their own standalone argus watching it.  So lets start 
with hardware. I have two identical sensor boxes which consist of Tyan Thunder 
motherboards with dual 1.4 gig Athelon processors and 500 megs or RAM. Each of 
those has 2 SysKonnect sx fibre gig cards in the 64 bit PCI slots, a 3c905b in 
a slot for connectivity and dual on board 3c905b used for monitoring the 
commodity link. This particular motherboard was bought because a benchmark
from a US university achieved ~950 megabits per second (which we have been able
to regularly duplicate I was able to get up to about 1.6 gig FDX during tests,
I think the limitation is likely memory bandwith but may be interrupt load) 
with this motherboard/card combo. It isn't cheap, but it is capable of wire 
speed and I highly recommend them (DAC cards are probably even better, but 
more expensive yet :-)).
	The original intent was one box for C4 and one box for commodity, but 
traffic has been light enough so far that a single sensor box does both 
commodity and C4 at the moment. The links are both tapped (a netoptics fibre 
tap for C4, and a finsair (nee shomoti) 10/100 Century tap for commodity about 
to be replaced by dual 4 port netoptics regen taps) and feed the SysKonnect 
cards and the 2 on board NICs for FDX operation on the first sensor machine. 
That machine runs argus_bpf (on FreeBSD 4.10-RELEASE) writing to a socket (no 
local archiving to disk for performance reasons) and basically loafs :-). 
Beside it is a P3 600 Meg machine with 750 Megs of ram and a couple of 3c905B 
cards. One card connects via a crossover cable to the sensor box 3c905b in the 
expansion slot (so external network events don't affect argus data capture). 
This machine has an 80 gig IDE disk and is the primary archive host. A copy of 
ra (daemonized by a quick and dirty perl script) listens on the crossover 
cable interface and argusarchive runs out of cron to cycle the logs every 
hour. On both machines. perl scripts running out of cron check that the 
appropriate argus tasks are around and will kill and restart them if they
look to have gone west in a variety of interesting ways. On the archive host
there are more perl scripts (that argusarchive triggers) which establish an
ssh connection to third machine in my office with 400 gigs of IDE disk and
transfers (reasonably reliably, although it still needs some more work) the 
archive files to a second archive for long term storage (backed up to our 
robot tape library for archival storage). All the perl scripts that post 
process the data for reports run on this third machine so screw ups or running
the machine out of memory don't affect the production sensor/archive pair. As
well because the post processing machine is a fibre link (and some 20 KM) from
the sensors, should the link go down, the archive machine beside the sensor
will happily puff its cheeks up with the data and then spit it all out 
(assuming it hasn't run out of disk which is fairly unlikely) when the link
comes back up.
	As of last weekend my other Athelon box is connected kluged (via our
multiport monitoring fibre tap at the moment until I get a dedicated second 
tap or the regen taps come) in to the gig link for the grid computing network.
That sensor is identical (except without 100 interfaces) to the one downtown
but its ra archive is running on my post processing box rather than a dedicated
archiving box because it is mostly experimental. I have hacked argusarchive 
so that it takes an instance name and stores the data for all the various 
instances in their own archives, and the post processing perl scripts all
deal with instance names (and use locking to serialize post processing to 
avoid blowing the machines out of memory most of the time). The initial 
versions of all the scripts are available via anon ftp at  ftp.sfu.ca  in 
/pub/unix/argus/argus.traffic.perl.tar.gz
Once I get more bugs and performance issues resolved I'll do a further release
(the interhost transfer code isn't currently there for instance)..
	As noted earlier every hour when the archive cycles a perl script 
is kicked off on the post processing host which aquires traffic and scan
information from the last hour's data and writes it to a spool directory.
Each morning at 6 AM another perl script kicks off and postprocessess the 
hourly summary files in to a set of reports covering the entire day (currently
not including port scans which are still being worked on) for commodity, C4 
and now Westgrid argus instances. It keeps the last 7 days of reports around
unless I choose to move them for some reason and are entirely expendable since
thay can be easily recreated from the argus archive data.
	The Surrey campus which is currently on the end of an independent
10 meg service runs an all in one argus instance on a P3 1 gig PC. Because the
link is only 10 full we haven't separated sensor and archive host so the entire
process including post processing runs on that one box (so far without apparant
problem). When the dark fibre appears, it will move behind our border router
and become part of the main argus instance and the local one will either go
away or be upgraded to a split sensor / archiver at gig (because the link will
go to gig) if they choose and still want to see whats going on on their link
out.

Peter Van Epp / Operations and Technical Support 
Simon Fraser University, Burnaby, B.C. Canada



More information about the argus mailing list