Switches, Hubs, and Modems
1747993 Members
5034 Online
108756 Solutions
New Discussion юеВ

Re: 16-port 10/100/1000 Module (J4907A) failure

 
Les Ligetfalvy
Esteemed Contributor

Re: 16-port 10/100/1000 Module (J4907A) failure

Update:
The module that faulted is a replacement module from HP. The original module and chassis was returned to HP. The replacement chassis took a few weeks to arrive, so I ended up taking my spare off the shelf, placing the replacement on the shelf as a spare when it finally arrived.

So, to summarize... I have had three failures, two of which HP has replaced. All three chassis were close to each other in serial number (I had ordered seven at one time).
They end in:
08S - current failure
090 - replaced
099 - replaced

The remaining siblings are:
09C
0BY
0KJ
0KM

The units that failed were populated with six or more J4907A modules. The remaining units have fewer modules.
Les Ligetfalvy
Esteemed Contributor

Re: 16-port 10/100/1000 Module (J4907A) failure

blogging...
Tried running some diags but the reboot into diag mode cleared the fault. :(

I removed the switch from its service location and hauled it back to my test bench. I tried 21 times to fault it by cycling RPS2 but to no avail.

One thing I have noticed is that in every case where there was a module fault, the switch had six or eight modules. and that the fault followed closely after a cycling or RPS2. I inserted two more modules to stack the odds in my favor, taking the total to eight, and cycled RPS2 13 more times.

I have the faulting 5308xl connected via module B port 7 to my 2524 on port 1. My WUG syslog shows this port #1 toggling off and on several times one minute later for the next minute. This I believe is when the switch module C faulted/crashed/rebooted.

I moved all the modules to another chassis (maintaining slot for slot) with different RPSs and will continue the testing.
Les Ligetfalvy
Esteemed Contributor

Re: 16-port 10/100/1000 Module (J4907A) failure

I also got the module to fault in another chassis with different RPSs. Next I will put a another identical module in slot C and try to fault it.
Les Ligetfalvy
Esteemed Contributor

Re: 16-port 10/100/1000 Module (J4907A) failure

Well... I must say that I am not impressed. Eight days after reporting this (10/21/04), I am still waiting for a replacement and if (big IF) it were to arrive on Monday, it will be eleven. I am also not impressed that it took until this week for HP to start looking at the original two faulted units that I sent out on 9/21/04.

Now if i started counting when the first fault was reported on 8/25/04... my $100,000 that is tied up in these switches that I have yet to put in production...

If I get chastized for saying "the emperor has no clothes" then so be it. I have run out of patience! Like Dan said in another thread, have I paid $100,000 for you, folks at HP, to sleep well?
Les Ligetfalvy
Esteemed Contributor

Re: 16-port 10/100/1000 Module (J4907A) failure

Well... I may as well update this blog. I got word on the results of the forensic testing commissioned by HP Division, and was told that there was an off-spec component. There is no plan to recall any other modules so I assume any possible future failures would be processed as regular warranty replacements through normal channels.
Les Ligetfalvy
Esteemed Contributor

Re: 16-port 10/100/1000 Module (J4907A) failure

Yesterday I had yet another module fail. This module had no connections to it so one cannot attribute the falure to ouside forces like ESD. I have had so many issues with these switches that the deployment has halted. This particular switch presently serves only to perpetuate the mesh. It has three 16-port gig modules that were to serve workstations on our production floor but for obvious reasons, I halted the migration. Beside this HP switch, I still have my Nortel Baystack switches serving the production wrapline.

Everytime my confidence starts to grow to the point I consider resuming the migration, along comes another speedbump. How do other people deploy this product in a 24/7 production environment and sleep at night (or do they)?
William_169
Occasional Advisor

Re: 16-port 10/100/1000 Module (J4907A) failure

Hi Les,

Is the problems with the 5308XL solved?
Is this a bad switch?
Do others have the same issues as yours?
What is the equivalent in Cisco?
Thanks.

William
william@gndt.com
Les Ligetfalvy
Esteemed Contributor

Re: 16-port 10/100/1000 Module (J4907A) failure

William,
I really don't know how to answer your question. If I really said how I felt right now, my comments would be censored. See also my other posts:
http://forums1.itrc.hp.com/service/forums/pageList.do?userId=CA1185869&admit=716493758+1113785918617+28353475&listType=question&forumId=1

You will just have to draw your own conclusions.
William_169
Occasional Advisor

Re: 16-port 10/100/1000 Module (J4907A) failure

We have several 5372XL with J4820A and J4907A. So far, we didn't have many problems. The problem we had was the J4820A was the "older" version and HP replaced them all quickly.
Do you think your problems are related to the meshing?
Did you have such problems with only the 10/100/1000 J4907A modules?
What if you only use the 10/100 J4820A?
Thanks.

Les Ligetfalvy
Esteemed Contributor

Re: 16-port 10/100/1000 Module (J4907A) failure

William,
The J4907A modules are the only copper modules that I use. I purchased 8 of the J8167A bundles (HP ProCurve Switch 5308XL-48G 8-slot chassis with three 16-port modules pre-installed) so that gives me 24 of these gig modules and I have had 4 of them go bad already. HP says that statistically I should not be experiencing this many failures. With these odds, I should have put the $100,000 on lottery tickets!

I have several J4852A
12 port 100FX modules that I use to connect to my Cisco core switch and my Nortel Baystack edge switches. I already have 100 meg copper on my Nortels so there is no reason for me to put in HP 100 meg copper modules.

As to whether these modules are reponsible for my mesh problem, that is the 100,000 dollar question that even a team of engineers at HP Division cannot answer.