Difference between revisions of "Topics for the 2015 Software Review"
From GlueXWiki
Line 7: | Line 7: | ||
#** 4500 Concurrent Jobs at JLab | #** 4500 Concurrent Jobs at JLab | ||
#** Well under 0.1% failure rate | #** Well under 0.1% failure rate | ||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
#* DC3 - January/February 2015 | #* DC3 - January/February 2015 | ||
#** Read data in raw-event format from tape and produce to DST (REST) files. | #** Read data in raw-event format from tape and produce to DST (REST) files. | ||
Line 18: | Line 12: | ||
#** Run Multi-threaded jobs | #** Run Multi-threaded jobs | ||
#** Already doing full reprocessing of the Fall 2014 data from tape every two weeks. | #** Already doing full reprocessing of the Fall 2014 data from tape every two weeks. | ||
− | + | # Data Acquisition Successes - Running Fall 2014 | |
− | * | + | #* Exceeded the 300MB/s transfer to tape bandwidth of experiment. |
− | ** | + | #** ~500 million events. |
− | ** Event | + | #** 7000 files, 120TB of data |
− | ** | + | #* Most data were taken in full pulse mode of the Flash ADCs |
− | ** Handle corrupted evio data | + | #** Need to get final processing algorithms on the FPGAs in the FADCs |
+ | #** Need to clean raw data of massive unused headers. | ||
+ | #* Event Rates of 2KHz for full experiment, much higher for individual components. | ||
+ | #** Need to move to block mode. | ||
+ | #** Need to move to FPGA processing to compress data. | ||
+ | #* Full DAQ chain to local raid disk, transfer to tape, and automatic processing from tape. | ||
+ | #* Robustness issues with the system | ||
+ | #** Handle corrupted evio data | ||
+ | #** Problems with some FADCs getting out of sync. | ||
+ | #* Stealth Online Data Challenge | ||
* Revist data spreadsheet from a year ago | * Revist data spreadsheet from a year ago | ||
* offline monitoring | * offline monitoring |
Revision as of 13:19, 17 January 2015
- Report on Successful Data Challenges
- DC1 - December 2012/ January 2013
- 5 billion Events - OSG, JLab, CMU
- 1200 Concurrent Jobs at JLab.
- DC2 - March/April 2014
- 10 billion events with EM backgrounds included - OSG, JLab, MIT, CMU, FSU
- 4500 Concurrent Jobs at JLab
- Well under 0.1% failure rate
- DC3 - January/February 2015
- Read data in raw-event format from tape and produce to DST (REST) files.
- Load up as many JLab cores as possible.
- Run Multi-threaded jobs
- Already doing full reprocessing of the Fall 2014 data from tape every two weeks.
- DC1 - December 2012/ January 2013
- Data Acquisition Successes - Running Fall 2014
- Exceeded the 300MB/s transfer to tape bandwidth of experiment.
- ~500 million events.
- 7000 files, 120TB of data
- Most data were taken in full pulse mode of the Flash ADCs
- Need to get final processing algorithms on the FPGAs in the FADCs
- Need to clean raw data of massive unused headers.
- Event Rates of 2KHz for full experiment, much higher for individual components.
- Need to move to block mode.
- Need to move to FPGA processing to compress data.
- Full DAQ chain to local raid disk, transfer to tape, and automatic processing from tape.
- Robustness issues with the system
- Handle corrupted evio data
- Problems with some FADCs getting out of sync.
- Stealth Online Data Challenge
- Exceeded the 300MB/s transfer to tape bandwidth of experiment.
- Revist data spreadsheet from a year ago
- offline monitoring
- browser
- analyze data as it appears on the silo
- reconstruction results
- calibration committee
- bi-weekly meeting
- preliminary list of constants compiled
- calibration still needs to be regularized
- calibration database training
- CCDB successes
- command line interface
- SQLite form of database
- analysis results
- electron identification in the FCAL.
- pi0 peak
- proton id with tof
- rho meson in pi+ pi-
- omega meson in pi+ pi- pi0
- data transfer to CMU via globus Online
- data management: event store, etc.
- stealth data challenge on real data