Closed Bug 427820 Opened 14 years ago Closed 14 years ago

Run a script to collect data for bug 423968


( Graveyard :: Server Operations, task)

Not set


(Not tracked)



(Reporter: benjamin, Assigned: aravind)



I've written a little script that will format the data kkovash wants into CSV format for him.

I've tested it on khan against the socorro development DB, and I'd like to get it run now against production, and the data uploaded somewhere:

the script is khan:/home/bsmedberg/
You'll need to edit the header which has username/password information.
It will need to run on a host which has psycopg2 installed, which should be any of the socorro-capable VMs.

Just run it with "python > data.csv". It dumps the reports and modules from the last week, which shouldn't be a terrible DB load since it can use the index pretty effectively.
That discussion is misplaced, should be in bug 423968.
Blocks: 423968
No longer blocks: 412605
Summary: Run a script to collect data for bug 412605 → Run a script to collect data for bug 423968
No longer blocks: 412605
Done.  Its a 70MB output file.
Closed: 14 years ago
Resolution: --- → FIXED
Benjamin: it looks like I was a bit too ambitious in the volume of data I was requesting (a laptop won't be able to play around with it).  can I ask for an adjustment to script?

can you pull a smaller date range (e.g., one day rather than one week)?  the file had something like 250K+ rows/crashes.  if we can cut that down to something less than 100K, that'd be great.

also, can you omit the filename versions (i.e., just include the filename fields).  it looked like there were something like 1,200 fields/variables for each crash... if we can cut this down as well, the data would be more workable.
Sure. We don't need to run the script again, I'll just text-process it down using python.
Ken, has what you need
Product: → Graveyard
You need to log in before you can comment on or make changes to this bug.