My suspicions about the devolo kit being on the slow side have turned out to be partly true, but they were really the least of my problems. I took an hour or two to revive an ancient laptop (from when MMX instructions were something to brag about) to test different points around the network, configurations, etc. As old as the laptop is, it's quite capable of doing a few things, and it's reliable.
What I couldn't figure out was why my new PC was only able to move things over the LAN at 0.3Mb/s! It's almost modem type of speeds. Because of the way my flat is laid out it was awkward to isolate the new machine and find out the exact cause. I (literally) dusted off an ancient laptop, and pressd it into service.
After continuously rewiring most of my flat and retesting, it became clear that my new PC, all on it's lonesome, was the cause of the glacial network throughput. After some web searches, auto-scaling on vista was pointed at - no luck (didn't think so, there weren't routers in this path).
More web searches suggested NAM wasn't good. It took awhile longer to discover that NAM is Nvidia's "Network Access Manager", and it's rubbish. I uninstalled it.
Without the devolo kit or NAM in the loop, network throughput jumped up towards the expected speeds (100Mb/s). I've put the devolo kit back into it's original places, and it's not too quick - ~11Mb/s; but my best option right now (and thirty times faster than NAM was forcing me into).
During testing I found that the devolo kit, when installed in sensible places (no power strips, etc.) will get close to it's theoretical speeds. Happy days - if you have power sockets free in the right places. I'll make do.
The bits of the IT world that apply to me right now. Blogged in the hope that someone (even me) will find them useful.
Wednesday, February 20, 2008
Devolo not-so-bad
Labels:
Devolo,
homeplug,
lan troubleshooting,
nam,
network access manager,
nvidia
Monday, February 18, 2008
Perl threads :-(
I chucked in an entry for the Computer languages shootout game for the Perl implementation of a thread-ring. I've a few observations:
I'm going to see if libevent holds any hope for me. If not, then maybe a mass of file handles, and select statements...
- My application relies on semaphores that unblock themselves quickly. On average, there's two threads running (rather than blocking on a semaphore) at a time, but my dual core PC is only 60-80% utilised. Something in the cond_wait implementation is sleeping.
- Try as I might, semaphores were the only IPC mechanism I could get to work in a way that appeared "correct" (rather than spurious errors about shared or locked, or whatever). This may be a reflection on the workman rather than his tools.
- Getting any cross-platform, efficient, thread-safe, IPC across muliple versions of Perl, is like pulling teeth. I've got a blended semaphore-based implementation ready to go. By blended, I mean it uses (inefficient) semaphores for the tricky first lap around the ring (ensuring that all initialisation is done correctly), and then move over to "something" immediate for message passing. I need a working message passing implementation to slot in now.
- As the shootout game demonstrates, Perl threads are not even close to universally available in any sense of the term.
I'm going to see if libevent holds any hope for me. If not, then maybe a mass of file handles, and select statements...
Labels:
Debian,
multi-threaded,
Perl,
progamming,
semaphores,
threads
Googlewhack!
Saturday, February 16, 2008
Devolo not-so-highspeed
Update on this item
The devolo "highspeed" homeplug gizmos are pretty good. There was a little buzz of excitement when the idea of power cables being - well, just power cables - suddenly lurched to the left by some distance and the idea of "just fat, slow cat5" snuck in next door.
The 85Mb/s claimed by Devolo isn't exactly electrifying (heh?), but that's ok, I only recently moved out my last 10Mb PC (100MBit still feels so fast). Anyway, it seems to move standard resolution movies (about 700MB or so) to the XBMC ok, and that's the one my girlfriend will be upset about. What the Devolo kit doesn't seem to is actually give me decent network throughput.
I did some testing using netcat - 6Mb/s! Ok, so the power/network isn't ideal (old wiring, too many power strips, etc.), but the "informer" app says I'm getting about 50Mb/s, so long as there's no actual traffic. If I actually ask the network to carry some traffic (as with netcat), the informer app says the devolo kit drops to 20Mb/s, while I'm only getting 6Mb useful throughput.
If I can improve the wiring setup (attach the devolo kit to it's own sockets rather than on power strips), I'll retest. So far though, in my little flat, with too few power sockets, in odd locations, the results aren't too encouraging. Do I currently feel like shelling out for the newer 200Mb/s model? Not really.
Maybe 802.11n is the next step - even with the thick walls here...
The devolo "highspeed" homeplug gizmos are pretty good. There was a little buzz of excitement when the idea of power cables being - well, just power cables - suddenly lurched to the left by some distance and the idea of "just fat, slow cat5" snuck in next door.
The 85Mb/s claimed by Devolo isn't exactly electrifying (heh?), but that's ok, I only recently moved out my last 10Mb PC (100MBit still feels so fast). Anyway, it seems to move standard resolution movies (about 700MB or so) to the XBMC ok, and that's the one my girlfriend will be upset about. What the Devolo kit doesn't seem to is actually give me decent network throughput.
I did some testing using netcat - 6Mb/s! Ok, so the power/network isn't ideal (old wiring, too many power strips, etc.), but the "informer" app says I'm getting about 50Mb/s, so long as there's no actual traffic. If I actually ask the network to carry some traffic (as with netcat), the informer app says the devolo kit drops to 20Mb/s, while I'm only getting 6Mb useful throughput.
If I can improve the wiring setup (attach the devolo kit to it's own sockets rather than on power strips), I'll retest. So far though, in my little flat, with too few power sockets, in odd locations, the results aren't too encouraging. Do I currently feel like shelling out for the newer 200Mb/s model? Not really.
Maybe 802.11n is the next step - even with the thick walls here...
Friday, February 15, 2008
SOA monitoring and discovery system
SOA is a great idea, taking chunks from EAI lessons learned, and generalising them. It occurs to me that tools for building maps of service invocations is necessary and "good". The question becomes, fairly quickly, "how?".
Asking developers and designers to be good citizens and update spreadsheets (or more sophisticated tools) doesn't work, because over a little time, the manual update system breaks down for lots of reasons (not least laziness). I suggest the only real option is to automate discovery of service invocations by watching the service interfaces themselves.
Point-to-point transports (usually TCP-based, like HTTP) make capturing traffic quite awkward, but the content can usually be extracted as it will pass through so many network elements (including the service's host machine) where a wire tap can be taken. If the communication is encrypted in-flight (as is expected for sensitive data), then an application level wiretap will be required.
BEA's AquaLogic took a very active role in this monitoring when I grappled with it; essentially as a proxy in the whole process. Maybe they made less intrusive options available in version 3... My first outing with AquaLogic didn't encourage a second try. All-in-all AquaLogic just seemed an order of magnitude too big and clunky.
I favour a distributed system of agents suitable for capturing each service invocation. Each agent could feed (largely unprocessed) data back to a cluster of analysis machines. At the end, I'd hope to see a bunch of query/display/integration tools to leverage (what would be in my experience) the only up-to-the-minute source of information about how the organisation actually hangs together.
A JMS wiretap (vendor specific, or fall-back to browsing the queues), could be implemented easily enough. For HTTP snooping, tcpdump's ok. For HTTPS, web server modules would be required (Apache, IIS?, ligHTTP?).
The analysis? Another day...
Asking developers and designers to be good citizens and update spreadsheets (or more sophisticated tools) doesn't work, because over a little time, the manual update system breaks down for lots of reasons (not least laziness). I suggest the only real option is to automate discovery of service invocations by watching the service interfaces themselves.
Point-to-point transports (usually TCP-based, like HTTP) make capturing traffic quite awkward, but the content can usually be extracted as it will pass through so many network elements (including the service's host machine) where a wire tap can be taken. If the communication is encrypted in-flight (as is expected for sensitive data), then an application level wiretap will be required.
BEA's AquaLogic took a very active role in this monitoring when I grappled with it; essentially as a proxy in the whole process. Maybe they made less intrusive options available in version 3... My first outing with AquaLogic didn't encourage a second try. All-in-all AquaLogic just seemed an order of magnitude too big and clunky.
I favour a distributed system of agents suitable for capturing each service invocation. Each agent could feed (largely unprocessed) data back to a cluster of analysis machines. At the end, I'd hope to see a bunch of query/display/integration tools to leverage (what would be in my experience) the only up-to-the-minute source of information about how the organisation actually hangs together.
A JMS wiretap (vendor specific, or fall-back to browsing the queues), could be implemented easily enough. For HTTP snooping, tcpdump's ok. For HTTPS, web server modules would be required (Apache, IIS?, ligHTTP?).
The analysis? Another day...
Labels:
AquaLogic,
architecture,
BEA,
EAI,
HTTP monitoring,
integration,
JMS,
service discovery,
SOA
Subscribe to:
Posts (Atom)