Closed Bug 1236021 Opened 10 years ago Closed 10 years ago

brasstack is throwing 502

Categories

(Infrastructure & Operations Graveyard :: WebOps: Other, task)

task
Not set
normal

Tracking

(Not tracked)

RESOLVED FIXED

People

(Reporter: Usul, Unassigned)

Details

(Whiteboard: [nsm] [kanban:https://webops.kanbanize.com/ctrl_board/2/2386] )

Whiteboard: [kanban:https://webops.kanbanize.com/ctrl_board/2/2384]
Whiteboard: [kanban:https://webops.kanbanize.com/ctrl_board/2/2384]
Whiteboard: [kanban:https://webops.kanbanize.com/ctrl_board/2/2386]
Whiteboard: [kanban:https://webops.kanbanize.com/ctrl_board/2/2386] → [nsm] [kanban:https://webops.kanbanize.com/ctrl_board/2/2386]
logon to server notice process for app not running on brasstacks1.dmz.sjc1.mozilla.com. Restarted it [root@brasstacks1.dmz.scl3 ~]# ps -ef | grep -i orangefactor root 32508 32481 0 13:54 pts/0 00:00:00 grep -i orangefactor [root@brasstacks1.dmz.scl3 ~]# /etc/init.d/orangefactor stop; /etc/init.d/orangefactor start; /etc/init.d/nginx reload stopping orangefactor [FAILED] starting orangefactorspawn-fcgi: child spawned successfully: PID: 32524 [ OK ] Reloading nginx: [ OK ] [root@brasstacks1.dmz.scl3 ~]# ps -ef | grep -i orangefactor webtools 32524 1 1 13:55 ? 00:00:06 /home/webtools/apps/orangefactor/bin/python /home/webtools/apps/orangefactor/src/orangefactor/server/woo_server.fcgi root 32568 32481 0 14:01 pts/0 00:00:00 grep -i orangefactor [root@brasstacks1.dmz.scl3 ~]# Site confirmed working now from Moc
Site's repaired, woot. Ashish, do we want to close a monitoring gap for comment 1?
Status: NEW → RESOLVED
Closed: 10 years ago
Flags: needinfo?(ashish)
Resolution: --- → FIXED
(In reply to Richard Soderberg [:atoll] from comment #2) > Site's repaired, woot. Ashish, do we want to close a monitoring gap for > comment 1? Host monitoring was removed in Bug 1231001.
Flags: needinfo?(ashish)
While I'm glad this was an easy fix, I'm troubled in light of 1231001#c2 which basically asks for all monitoring to be removed. :emorley, should we instruct MOC to refuse all participation on this server, or should we add a basic process check for the obvious fix above, or what's the correct plan forward here? It's an IT-managed server, so we're expected to manage it, and shutting off all monitoring wholesale seems guaranteed to make IT look bad for no good reason :(
Flags: needinfo?(emorley)
I don't believe it should be IT managed; it's effectively an a-team project server running an EOL webapp that we hope to replace with equivalent Treeherder functionality (and decom the box) in the next few months.
Flags: needinfo?(emorley)
Ah, interesting. Okay! We will keep that in mind when people report breakage in the future. I assume it's safe to try and perform the basic repair steps from comment 1 as desired, without any particular harm to it?
Yup :-)
Product: Infrastructure & Operations → Infrastructure & Operations Graveyard
You need to log in before you can comment on or make changes to this bug.