I had already included a NETtool in the original. This is a copy of the
link control. And operating system is 7.5 powerpatch 3. I'm also
including the error that shows in NETTOOL NMDUMP.
**********************************************************************
* WED, JAN 28, 2009, 11:42:51.8 AM NETXPORT(3) *
*--------------------------------------------------------------------*
* Event : BUFFER MANAGER *
* Entity : UDP *
* Internal Event : Buffer manager error *
* Log Class : Operator intervention/attention required *
* Port ID/PIN/KSO : $FFFFF65E *
* Location : 13 Parameter : $FFE900C9 *
* Info Section (hex): *
* 0000: 008D 0006 0007 000D FFFF F65E 0002 FFE9 *
* 0008: 00C9 0001 *
*--------------------------------------------------------------------*
* Port Message Frame: *
* Function Code : IPC, DATAGRAM SEND REQUEST *
* Reply Port ID : $00000000 *
* Subqueue Number : 7 Message Length : 26 *
* Reply Subqueue #: 7 Flow ID : $0000 *
* Data Section (hex): *
* 0000: 0007 001A 0000 0000 0007 050B 0000 0002 *
* 0008: 0000 020D 0001 DE28 B384 0087 AACC 0E50 *
* 0010: 0000 0024 0030 0030 0004 0050 0000 050D *
* 0018: 4184 EF28 *
**********************************************************************
linkcontrol @,a
Linkname: DTSLINK Linktype: PCI 100BT Linkstate: CONNECTED
Physical Path: 1/12/0/0
Current Station Address: 00-30-6E-06-01-E4
Default Station Address: 00-30-6E-06-01-E4
Current Multicast Addresses:
09-00-09-00-00-01 09-00-09-00-00-03 09-00-09-00-00-04
09-00-09-00-00-06
Transmit bytes 277420582 Receive bytes
96254205
Transmits 1623708 Receives unicast
1070001
Transmits no error 1623708 Receives broadcast
25269
Transmits dropped 0 Receives multicast
1606
Transmits deferred 0 Receives no error
1096549
Transmits 1 retry 0 Recv CRC error
0
Transmits >1 retry 0 Recv Maxsize error
0
Trans 16 collisions 0 Recv dropped: addr
327
Trans late collision 0 Recv dropped: buffer
0
Trans underruns 0 Recv dropped: descr
0
Carrier losses 0 Recv dropped: other
0
Trans jabber timeout 0 Recv watchdg timeout
0
Link disconnects 0 Recv collisions
0
Link speed 100 Recv overruns
0
Link duplex Full Link auto sensed
No
Link mode 100Base-TX Addon Secs since clear
35212
From: Craig Lalley [mailto:[log in to unmask]]
Sent: Thursday, January 29, 2009 10:01 AM
To: [log in to unmask]; Stanfield, Randy (Carrollton, TX)
Subject: Re: : Network Issue
Randy,
What version of MPE are you on?
Are you using a 100mb full duplex or 10mb half duplex card?
If you are experincing a network hang, that would be a good time to do
the
NETTOOLS.NET.SYS --> RESOURCE --> DISPLAY
and
LINKCONTROL @,a
I would also be helpful to know how the TCP/IP Parameters are set in
NMMGR.
Regards,
-Craig
--- On Thu, 1/29/09, Stanfield, Randy (Carrollton, TX)
<[log in to unmask]> wrote:
From: Stanfield, Randy (Carrollton, TX) <[log in to unmask]>
Subject: : Network Issue
To: [log in to unmask]
Date: Thursday, January 29, 2009, 7:45 AM
Network help
We've started using some QSDK
software to allow the HP3000 to become a
Webserver basicly. We've been using is pretty successfully I might add
but the volume is starting to increase and noticing some bottle necks
that look to be network related. Does anyone have a few minutes to maybe
look at our issue. I'm not sure what to change if anything to change in
the size of any of the buffer pools for the network. We had our OUTBOUND
Buf Pool on Loop max out yesterday and cause the Webservices to hang. We
increased that size to 512 about 2 weeks ago and that helped but didn't
want to keep changing without knowing what it affects.
[5]RESOURCE>>>dis
THU, JAN 29, 2009, 7:40:58 AM
Item Subsystem Name G/N Description Used High Max
____________________________________________________________________
1 NS XPORT CP_POOL_ (G) Control Buf Pool 0 1 300 :)
2 NS XPORT
1536___D (G) Inbound Buf Pool 42 60 512 :)
3 NS XPORT NET1____ (N) Outbnd Buf Pool 0 50 768 :)
4 NS XPORT LOOP____ (N) Outbnd Buf Pool 0 75 512 :)
5 NS XPORT UDP (G) GProt Msg Pool 0 N/A 512 :)
6 NS XPORT PXP (G) GProt Msg Pool 0 N/A 660 :)
7 NS XPORT PROBE (G) CM Prot Msg Pool 0 N/A 678 :)
8 NS XPORT IP_NI (G) IP-NI Msg Pool 0 N/A 2048 :)
9 NS XPORT IP_NI (G) IP-NI Msg Pool 0 N/A 2048 :)
10 NS XPORT (G) Node Name Cache 0 1 360 :)
11 NS XPORT (G) TCP Control pool 0 73 2048 :)
12 TELNET PTOD (G) Write Buf Pool 0 1 2048 :)
13 TELNET PTID (G) Outbnd Buf Pool 0 1 2048 :)
14 TELNET PTID (G) Inbound Buf Pool 1 2
2048 :)
15 TELNET PTOD (G) Negot Buf Pool 0 1 2048 :)
Randy Stanfield
Unisource - Dallas
972-982-3621
From: Fatula, Steve (Norcross, DAV)
So, we have a program WBGETPR that primarily calculates cost and price
on HP3000. It is called via network calls as a XML web service. So,
client pc or machine of some sort calls the HP3000, sends XML string in
a POST, and, gets reply back on the same socket. We have a LOT of these
types of services on the HP, and all of them seem to work well. Except,
this one has some odd behavior we'd like to understand better. The
program is different than the other HP services we have in that it seems
to return a far larger amount of data. This service not only prices one
item, it prices a list of items in a single call. There is a lot of data
returned for each item. So, when we get
a list of say 300 items to price
and cost, the reply data could be around 400K. This has been
horrendously slow.
So, we put some timers in the code. Where the process bogs down is in
the network WRITES from the HP to the client. The client can be an
HP3000, a PC running proprietary software, or, a PC running the OPENSTA
open source test tool, it made no difference. The data sent to the
client is divided up into chunks of at most 30,000 bytes. The HP call
used is IPCSEND. The timers around calls to IPCSEND showed elapsed run
times in the hundredth of a second or smaller for the first two IPCSEND
calls in a reply, then would spike, to 4, 8, or even 15 seconds for a
single IPCSEND call. Drastic change! This would cause the process to be
very slow needles to say. We never saw the first two chunks of 30,000
bytes slow, they were always almost instant. Always, the third one (and
some of the later ones)
was slow, and the time varied widely. This is on
a very fast development machine. We monitored some of the network
tables, and didn't find any that maxed out. So, not sure where to look
to find out why this might occur.
As an experiment, we changed our maximum size to 10,000 bytes in the
IPCSEND calls. This resulted in no times exceeding .even 1 second, I
believe the largest we saw as .6 and that was with a bunch of them
running at the same time. But it was way faster than using 30,000 byte
IPCSEND calls.
The question is, why! What can be causing this behavior, and is there
maybe a more optimal number we can use other than 10,000. We'd like to
understand what the issue may be, it would seem 30,000 would be more
efficient since there are less calls, but that is not what we see.
Randy, you can add any details about the hardware, and, let me know if
you think I am missing
anything here.
Steve Fatula
5 Diamond IT Consulting
214-592-4230 Voice Ext 151
206-984-9737 Fax
* To join/leave the list, search archives, change list settings, *
* etc., please visit http://raven.utc.edu/archives/hp3000-l.html *
* To join/leave the list, search archives, change list settings, *
* etc., please visit http://raven.utc.edu/archives/hp3000-l.html *
|