#ifndef MARS_DrsCalib #define MARS_DrsCalib #include // fabs #include // errno #ifndef MARS_fits #include "fits.h" #endif #ifndef MARS_ofits #include "ofits.h" #endif #ifdef __MARS__ #include "MTime.h" #endif class DrsCalibrate { protected: uint64_t fNumEntries; size_t fNumSamples; size_t fNumChannels; std::vector fSum; std::vector fSum2; public: DrsCalibrate() : fNumEntries(0), fNumSamples(0), fNumChannels(0) { } void Reset() { fNumEntries = 0; fNumSamples = 0; fNumChannels = 0; fSum.clear(); fSum2.clear(); } void InitSize(uint16_t channels, uint16_t samples) { fNumChannels = channels; fNumSamples = samples; fSum.resize(samples*channels); fSum2.resize(samples*channels); } void AddRel(const int16_t *val, const int16_t *start) { for (size_t ch=0; ch1439 ? ((ch-1440)*9+8)*1024 : ch*1024; for (size_t i=0; i=roi) return 0; double step = 0; // before double rms = 0; // before int cnt = 0; // Exclude TM channel for (int p=0; p<8; p++) { const size_t hw = ch0+p; const size_t sw = (map?map[hw]:hw)*roi + pos; const double diff = vec[sw]-vec[sw-1]; step += diff; rms += (vec[sw]-vec[sw-1])*(vec[sw]-vec[sw-1]); cnt++; } return cnt==0 ? 0 : step/cnt; } static void SubtractStep(const size_t ch0, const double avg, float *vec, int16_t roi, int32_t pos, const uint16_t *map=NULL) { if (pos==0 || pos>=roi) return; const int begin = avg>0 ? pos : 0; const int end = avg>0 ? roi : pos; const double sub = fabs(avg); for (int p=0; p<9; p++) { for (int j=begin; j::iterator beg, const std::vector::iterator end) { Step rc; for (auto it=beg; it!=end; it++) { rc.pos += it->pos; rc.avg += it->avg; rc.rms += it->avg*it->avg; } rc.cnt = end-beg; rc.pos /= rc.cnt; rc.avg /= rc.cnt; rc.rms /= rc.cnt; rc.rms = sqrt(rc.rms-rc.avg*rc.avg); return rc; } static Step CorrectStep(float *vec, uint16_t nch, uint16_t roi, const int16_t *prev, const int16_t *start, const int16_t offset, const uint16_t *map=NULL) { std::vector list; list.reserve(nch); // Fill steps into array // Exclude broken pixels? // Remove maximum and minimum patches (4max and 4min)? for (size_t ch=0; chavg << std::endl; const size_t skip = list.size()/10; rc = AverageSteps(list.begin()+skip, list.begin()+list.size()-skip); // std::cout << " A1=" << rc.avg << " rms=" << rc.rms << std::endl; } for (size_t ch=0; ch25 && p[i]-p[i+1]>25) { p[i] = (p[i-1]+p[i+1])/2; } if (p[i]-p[i-1]>22 && fabs(p[i]-p[i+1])<4 && p[i+1]-p[i+2]>22) { p[i] = (p[i-1]+p[i+2])/2; p[i+1] = p[i]; } } } } static void RemoveSpikes2(float *vec, uint32_t roi)//from Werner { if (roi<4) return; for (size_t ch=0; ch<1440; ch++) { float *p = vec + ch*roi; std::vector Ameas(p, p+roi); std::vector diff(roi); for (size_t i=1; i N1mean(roi); //for (size_t i=1; i 10) { p[i+1]= (Ameas[i+3] - Ameas[i])/3 + Ameas[i]; p[i+2]= 2*(Ameas[i+3] - Ameas[i])/3 + Ameas[i]; i += 3; continue; } if ( (diff[i+1]<-diff[i]*fract*2) && (diff[i+2]>10) ) { p[i+1] = (Ameas[i]+Ameas[i+2])/2; diff[i+2] = (p[i+1] + Ameas[i+3])/2 - Ameas[i+2]; i += 2; } // const float x = Ameas[i] - N1mean[i]; // if (x > -5.) // continue; // if (Ameas[i+2] - (Ameas[i] + Ameas[i+3])/2. > 10.) // { // p[i+1]= (Ameas[i+3] - Ameas[i])/3 + Ameas[i]; // p[i+2]= 2*(Ameas[i+3] - Ameas[i])/3 + Ameas[i]; // i += 3; // continue; // } // const float xp = Ameas[i+1] - N1mean[i+1]; // const float xpp = Ameas[i+2] - N1mean[i+2]; // if ( (xp > -2.*x*fract) && (xpp < -10.) ) // { // p[i+1] = N1mean[i+1]; // N1mean[i+2] = Ameas[i+1] - Ameas[i+3]/2; // // i += 2; // } } } } static void RemoveSpikes3(float *vec, uint32_t roi)//from Werner { const float SingleCandidateTHR = -10.; const float DoubleCandidateTHR = -5.; const std::vector src(vec, vec+roi); std::vector diff(roi); for (size_t i=1; i=DoubleCandidateTHR) continue; //bool checkDouble = false; // a single spike candidate if (diff[i] -1.6*diff[i]) { vec[i+1] = (src[i] + src[i+2]) / 2; i += 2; /*** NEW ***/ continue; /*** NEW ***/ } /* else { // do nothing - not really a single spike, // but check if it is a double checkDouble = true; }*/ } // a double spike candidate //if (diff[i]>DoubleCandidateTHR || checkDouble == 1) { // check the consistency with a double spike if ((diff[i+1] > -DoubleCandidateTHR) && (diff[i+2] > -DoubleCandidateTHR)) { vec[i+1] = (src[i+3] - src[i])/3 + src[i]; vec[i+2] = 2*(src[i+3] - src[i])/3 + src[i]; //vec[i] = (src[i-1] + src[i+2]) / 2.; //vec[i+1] = (src[i-1] + src[i+2]) / 2.; //do not care about the next sample it was the spike i += 3; } } } } static void SlidingAverage(float *const vec, const uint32_t roi, const uint16_t w) { if (w==0 || w>roi) return; for (float *pix=vec; pix,std::vector > GetSampleStats() const { if (fNumEntries==0) return make_pair(std::vector(),std::vector()); std::vector mean(fSum.size()); std::vector error(fSum.size()); std::vector::const_iterator it = fSum.begin(); std::vector::const_iterator i2 = fSum2.begin(); std::vector::iterator im = mean.begin(); std::vector::iterator ie = error.begin(); while (it!=fSum.end()) { *im = /*cnt ... mean /= fNumEntries; error = sqrt(error/fNumEntries - mean*mean); */ return make_pair(mean, error); } void GetSampleStats(float *ptr, float scale) const { const size_t sz = fNumSamples*fNumChannels; if (fNumEntries==0) { memset(ptr, 0, sizeof(float)*sz*2); return; } std::vector::const_iterator it = fSum.begin(); std::vector::const_iterator i2 = fSum2.begin(); while (it!=fSum.end()) { *ptr = scale*double(*it)/fNumEntries; *(ptr+sz) = scale*sqrt(double(*i2*int64_t(fNumEntries) - *it * *it))/fNumEntries; ptr++; it++; i2++; } } static double GetPixelStats(float *ptr, const float *data, uint16_t roi) { if (roi==0) return -1; const int beg = roi>10 ? 10 : 0; double max = 0; for (int i=0; i<1440; i++) { const float *vec = data+i*roi; int pos = beg; double sum = vec[beg]; double sum2 = vec[beg]*vec[beg]; for (int j=beg+1; jvec[pos]) pos = j; } sum /= roi-beg; sum2 /= roi-beg; if (vec[pos]>0) max = vec[pos]; *(ptr+0*1440+i) = sum; *(ptr+1*1440+i) = sqrt(sum2 - sum * sum); *(ptr+2*1440+i) = vec[pos]; *(ptr+3*1440+i) = pos; } return max; } static void GetPixelMax(float *max, const float *data, uint16_t roi, int32_t first, int32_t last) { if (roi==0 || first<0 || last<0 || first>=roi || last>=roi || last*pmax) pmax = ptr; max[i] = *pmax; } } const std::vector &GetSum() const { return fSum; } uint64_t GetNumEntries() const { return fNumEntries; } }; class DrsCalibrateTime { public: uint64_t fNumEntries; size_t fNumSamples; size_t fNumChannels; std::vector> fStat; public: DrsCalibrateTime() : fNumEntries(0), fNumSamples(0), fNumChannels(0) { InitSize(160, 1024); } DrsCalibrateTime(const DrsCalibrateTime &p) : fNumEntries(p.fNumEntries), fNumSamples(p.fNumSamples), fNumChannels(p.fNumChannels), fStat(p.fStat) { } virtual ~DrsCalibrateTime() { } double Sum(uint32_t i) const { return fStat[i].first; } double W(uint32_t i) const { return fStat[i].second; } virtual void InitSize(uint16_t channels, uint16_t samples) { fNumChannels = channels; fNumSamples = samples; fNumEntries = 0; fStat.clear(); fStat.resize(samples*channels); } void AddT(const float *val, const int16_t *start, signed char edge=0) { if (fNumSamples!=1024 || fNumChannels!=160) return; // Rising or falling edge detection has the advantage that // we are much less sensitive to baseline shifts for (size_t ch=0; ch<160; ch++) { const size_t tm = ch*9+8; const int16_t spos = start[tm]; if (spos<0) continue; const size_t pos = ch*1024; double p_prev = 0; int32_t i_prev = -1; for (size_t i=0; i<1024-1; i++) { const size_t rel = tm*1024 + i; const float &v0 = val[rel]; //-avg; const float &v1 = val[rel+1];//-avg; // If edge is positive ignore all falling edges if (edge>0 && v0>0) continue; // If edge is negative ignore all falling edges if (edge<0 && v0<0) continue; // Check if there is a zero crossing if ((v0<0 && v1<0) || (v0>0 && v1>0)) continue; // Calculate the position p of the zero-crossing // within the interval [rel, rel+1] relative to rel // by linear interpolation. const double p = v0==v1 ? 0.5 : v0/(v0-v1); // If this was at least the second zero-crossing detected if (i_prev>=0) { // Calculate the distance l between the // current and the last zero-crossing const double l = i+p - (i_prev+p_prev); // By summation, the average length of each // cell is calculated. For the first and last // fraction of a cell, the fraction is applied // as a weight. const double w0 = 1-p_prev; fStat[pos+(spos+i_prev)%1024].first += w0*l; fStat[pos+(spos+i_prev)%1024].second += w0; for (size_t k=i_prev+1; ksecondfirst / it->second; num++; } avg /= num; for (auto it=beg; it!=end; it++) { if (it->second>=fNumEntries-0.5) continue; // { // result[i+1].first = *is2; // result[i+1].second = *iw2; // } // else // { it->first = avg*fNumEntries; it->second = fNumEntries; // } } } } DrsCalibrateTime GetComplete() const { DrsCalibrateTime rc(*this); rc.FillEmptyBins(); return rc; } void CalcResult() { for (int ch=0; ch<160; ch++) { const auto beg = fStat.begin() + ch*1024; const auto end = beg + 1024; // First calculate the average length s of a single // zero-crossing interval in the whole range [0;1023] // (which is identical to the/ wavelength of the // calibration signal) double s = 0; double w = 0; for (auto it=beg; it!=end; it++) { s += it->first; w += it->second; } s /= w; // Dividing the average length s of the zero-crossing // interval in the range [0;1023] by the average length // in the interval [0;n] yields the relative size of // the interval in the range [0;n]. // // Example: // Average [0;1023]: 10.00 (global interval size in samples) // Average [0;512]: 8.00 (local interval size in samples) // // Globally, on average one interval is sampled by 10 samples. // In the sub-range [0;512] one interval is sampled on average // by 8 samples. // That means that the interval contains 64 periods, while // in the ideal case (each sample has the same length), it // should contain 51.2 periods. // So, the sampling position 512 corresponds to a time 640, // while in the ideal case with equally spaces samples, // it would correspond to a time 512. // // The offset (defined as 'ideal - real') is then calculated // as 512*(1-10/8) = -128, so that the time is calculated as // 'sampling position minus offset' // double sumw = 0; double sumv = 0; int n = 0; // Sums about many values are numerically less stable than // just sums over less. So we do the exercise from both sides. // First from the left for (auto it=beg; it!=end-512; it++, n++) { const double valv = it->first; const double valw = it->second; it->first = sumv>0 ? n*(1-s*sumw/sumv) : 0; sumv += valv; sumw += valw; } sumw = 0; sumv = 0; n = 1; // Second from the right for (auto it=end-1; it!=beg-1+512; it--, n++) { const double valv = it->first; const double valw = it->second; sumv += valv; sumw += valw; it->first = sumv>0 ? n*(s*sumw/sumv-1) : 0; } // A crosscheck has shown, that the values from the left // and right perfectly agree over the whole range. This means // the a calculation from just one side would be enough, but // doing it from both sides might still make the numerics // a bit more stable. } } DrsCalibrateTime GetResult() const { DrsCalibrateTime rc(*this); rc.CalcResult(); return rc; } double Offset(uint32_t ch, double pos) const { const auto p = fStat.begin() + ch*1024; const uint32_t f = floor(pos); const double v0 = p[f].first; const double v1 = p[(f+1)%1024].first; return v0 + fmod(pos, 1)*(v1-v0); } double Calib(uint32_t ch, double pos) const { return pos-Offset(ch, pos); } }; struct DrsCalibration { std::vector fOffset; std::vector fGain; std::vector fTrgOff; uint64_t fNumOffset; uint64_t fNumGain; uint64_t fNumTrgOff; uint32_t fStep; uint16_t fRoi; // Region of interest for trgoff uint16_t fNumTm; // Number of time marker channels in trgoff std::string fDateObs; std::string fDateRunBeg[3]; std::string fDateRunEnd[3]; std::string fDateEnd; // uint16_t fDAC[8]; DrsCalibration() : fOffset(1440*1024, 0), fGain(1440*1024, 4096), fTrgOff (1600*1024, 0), fNumOffset(1), fNumGain(2000), fNumTrgOff(1), fStep(0), fDateObs("1970-01-01T00:00:00"), fDateEnd("1970-01-01T00:00:00") { for (int i=0; i<3; i++) { fDateRunBeg[i] = "1970-01-01T00:00:00"; fDateRunEnd[i] = "1970-01-01T00:00:00"; } } DrsCalibration(const DrsCalibration &cpy) : fOffset(cpy.fOffset), fGain(cpy.fGain), fTrgOff(cpy.fTrgOff), fNumOffset(cpy.fNumOffset), fNumGain(cpy.fNumGain), fNumTrgOff(cpy.fNumTrgOff), fStep(cpy.fStep), fRoi(cpy.fRoi), fNumTm(cpy.fNumTm), fDateObs(cpy.fDateObs), fDateRunBeg(cpy.fDateRunBeg), fDateRunEnd(cpy.fDateRunEnd), fDateEnd(cpy.fDateEnd) { } void Clear() { // Default gain: // 0.575*[45590]*2.5V / 2^16 = 0.99999 V fOffset.assign(1440*1024, 0); fGain.assign (1440*1024, 4096); fTrgOff.assign(1600*1024, 0); fNumOffset = 1; fNumGain = 2000; fNumTrgOff = 1; fStep = 0; fDateObs = "1970-01-01T00:00:00"; fDateEnd = "1970-01-01T00:00:00"; for (int i=0; i<3; i++) { fDateRunBeg[i] = "1970-01-01T00:00:00"; fDateRunEnd[i] = "1970-01-01T00:00:00"; } } std::string ReadFitsImp(const std::string &str, std::vector &vec) { #ifndef __MARS__ std::fits file(str); #else fits file(str); #endif if (!file) { std::ostringstream msg; msg << "Could not open file '" << str << "': " << strerror(errno); return msg.str(); } if (file.GetStr("TELESCOP")!="FACT") { std::ostringstream msg; msg << "Reading '" << str << "' failed: Not a valid FACT file (TELESCOP not FACT in header)"; return msg.str(); } if (!file.HasKey("STEP")) { std::ostringstream msg; msg << "Reading '" << str << "' failed: Is not a DRS calib file (STEP not found in header)"; return msg.str(); } if (file.GetNumRows()!=1) { std::ostringstream msg; msg << "Reading '" << str << "' failed: Number of rows in table is not 1."; return msg.str(); } fStep = file.GetUInt("STEP"); fNumOffset = file.GetUInt("NBOFFSET"); fNumGain = file.GetUInt("NBGAIN"); fNumTrgOff = file.GetUInt("NBTRGOFF"); fRoi = file.GetUInt("NROI"); fNumTm = file.HasKey("NTM") ? file.GetUInt("NTM") : 0; if (file.HasKey("DATE-OBS")) fDateObs = file.GetStr("DATE-OBS"); if (file.HasKey("DATE-END")) fDateEnd = file.GetStr("DATE-END"); if (file.HasKey("RUN0-BEG")) fDateRunBeg[0]= file.GetStr("RUN0-BEG"); if (file.HasKey("RUN1-BEG")) fDateRunBeg[1]= file.GetStr("RUN1-BEG"); if (file.HasKey("RUN2-BEG")) fDateRunBeg[2]= file.GetStr("RUN2-BEG"); if (file.HasKey("RUN0-END")) fDateRunEnd[0]= file.GetStr("RUN0-END"); if (file.HasKey("RUN1-END")) fDateRunEnd[1]= file.GetStr("RUN1-END"); if (file.HasKey("RUN2-END")) fDateRunEnd[2]= file.GetStr("RUN2-END"); /* fDAC[0] = file.GetUInt("DAC_A"); fDAC[1] = file.GetUInt("DAC_B"); fDAC[4] = file.GetUInt("DAC_C"); */ vec.resize(1440*1024*4 + (1440+fNumTm)*fRoi*2 + 4); float *base = vec.data(); reinterpret_cast(base)[0] = fRoi; file.SetPtrAddress("RunNumberBaseline", base+1, 1); file.SetPtrAddress("RunNumberGain", base+2, 1); file.SetPtrAddress("RunNumberTriggerOffset", base+3, 1); file.SetPtrAddress("BaselineMean", base+4+0*1024*1440, 1024*1440); file.SetPtrAddress("BaselineRms", base+4+1*1024*1440, 1024*1440); file.SetPtrAddress("GainMean", base+4+2*1024*1440, 1024*1440); file.SetPtrAddress("GainRms", base+4+3*1024*1440, 1024*1440); file.SetPtrAddress("TriggerOffsetMean", base+4+4*1024*1440, fRoi*1440); file.SetPtrAddress("TriggerOffsetRms", base+4+4*1024*1440+fRoi*1440, fRoi*1440); if (fNumTm>0) { file.SetPtrAddress("TriggerOffsetTMMean", base+4+4*1024*1440+ 2*fRoi*1440, fRoi*fNumTm); file.SetPtrAddress("TriggerOffsetTMRms", base+4+4*1024*1440+ 2*fRoi*1440+ fRoi*fNumTm, fRoi*fNumTm); } if (!file.GetNextRow()) { std::ostringstream msg; msg << "Reading data from " << str << " failed."; return msg.str(); } /* fDAC[2] = fDAC[1]; fDAC[4] = fDAC[1]; fDAC[5] = fDAC[4]; fDAC[6] = fDAC[4]; fDAC[7] = fDAC[4]; */ fOffset.resize(1024*1440); fGain.resize(1024*1440); fTrgOff.resize(fRoi*(1440+fNumTm)); // Convert back to ADC counts: 256/125 = 4096/2000 // Convert back to sum (mean * num_entries) for (int i=0; i<1024*1440; i++) { fOffset[i] = fNumOffset *256*base[i+1024*1440*0+4]/125; fGain[i] = fNumOffset*fNumGain*256*base[i+1024*1440*2+4]/125; } for (int i=0; i &vec) const { const size_t n = 1440*1024*4 + 1440*fRoi*2 + fNumTm*fRoi*2 + 3; #ifndef __MARS__ std::ofits file(filename.c_str()); #else ofits file(filename.c_str()); #endif if (!file) { std::ostringstream msg; msg << "Could not open file '" << filename << "': " << strerror(errno); return msg.str(); } file.AddColumnInt("RunNumberBaseline"); file.AddColumnInt("RunNumberGain"); file.AddColumnInt("RunNumberTriggerOffset"); file.AddColumnFloat(1024*1440, "BaselineMean", "mV"); file.AddColumnFloat(1024*1440, "BaselineRms", "mV"); file.AddColumnFloat(1024*1440, "GainMean", "mV"); file.AddColumnFloat(1024*1440, "GainRms", "mV"); file.AddColumnFloat(fRoi*1440, "TriggerOffsetMean", "mV"); file.AddColumnFloat(fRoi*1440, "TriggerOffsetRms", "mV"); file.AddColumnFloat(fRoi*fNumTm, "TriggerOffsetTMMean", "mV"); file.AddColumnFloat(fRoi*fNumTm, "TriggerOffsetTMRms", "mV"); #ifdef __MARS__ const MTime now(-1); file.SetStr( "TELESCOP", "FACT", "Telescope that acquired this data"); file.SetStr( "PACKAGE", "MARS", "Package name"); file.SetStr( "VERSION", "1.0", "Package description"); //file.SetStr( "CREATOR", "root", "Program that wrote this file"); file.SetFloat("EXTREL", 1.0, "Release Number"); file.SetStr( "COMPILED", __DATE__" "__TIME__, "Compile time"); //file.SetStr( "REVISION", REVISION, "SVN revision"); file.SetStr( "ORIGIN", "FACT", "Institution that wrote the file"); file.SetStr( "DATE", now.GetStringFmt("%Y-%m-%dT%H:%M:%S").Data(), "File creation date"); file.SetInt( "NIGHT", now.GetNightAsInt(), "Night as int"); file.SetStr( "TIMESYS", "UTC", "Time system"); file.SetStr( "TIMEUNIT", "d", "Time given in days w.r.t. to MJDREF"); file.SetInt( "MJDREF", 40587, "MJD to UNIX time (seconds since 1970/1/1)"); #else DataWriteFits2::WriteDefaultKeys(file); #endif file.SetStr("DATE-OBS", fDateObs, "First event of whole DRS calibration"); file.SetStr("DATE-END", fDateEnd, "Last event of whole DRS calibration"); file.SetStr("RUN0-BEG", fDateRunBeg[0], "First event of run 0"); file.SetStr("RUN1-BEG", fDateRunBeg[1], "First event of run 1"); file.SetStr("RUN2-BEG", fDateRunBeg[2], "First event of run 2"); file.SetStr("RUN0-END", fDateRunEnd[0], "Last event of run 0"); file.SetStr("RUN1-END", fDateRunEnd[1], "Last event of run 1"); file.SetStr("RUN2-END", fDateRunEnd[2], "Last event of run 2"); file.SetInt("STEP", fStep, ""); file.SetInt("ADCRANGE", 2000, "Dynamic range of the ADC in mV"); file.SetInt("DACRANGE", 2500, "Dynamic range of the DAC in mV"); file.SetInt("ADC", 12, "Resolution of ADC in bits"); file.SetInt("DAC", 16, "Resolution of DAC in bits"); file.SetInt("NPIX", 1440, "Number of channels in the camera"); file.SetInt("NTM", fNumTm, "Number of time marker channels"); file.SetInt("NROI", fRoi, "Region of interest"); file.SetInt("NBOFFSET", fNumOffset, "Num of entries for offset calibration"); file.SetInt("NBGAIN", fNumGain/1953125, "Num of entries for gain calibration"); file.SetInt("NBTRGOFF", fNumTrgOff, "Num of entries for trigger offset calibration"); // file.WriteKeyNT("DAC_A", fData.fDAC[0], "Level of DAC 0 in DAC counts") || // file.WriteKeyNT("DAC_B", fData.fDAC[1], "Leval of DAC 1-3 in DAC counts") || // file.WriteKeyNT("DAC_C", fData.fDAC[4], "Leval of DAC 4-7 in DAC counts") || file.WriteTableHeader("DrsCalibration"); if (!file.WriteRow(vec.data()+1, n*sizeof(float))) { std::ostringstream msg; msg << "Writing data to " << filename << " failed."; return msg.str(); } return std::string(); } std::string ReadFitsImp(const std::string &str) { std::vector vec; return ReadFitsImp(str, vec); } bool IsValid() { return fStep>2; } bool Apply(float *vec, const int16_t *val, const int16_t *start, uint32_t roi) { if (roi!=fRoi) { for (size_t ch=0; ch<1440; ch++) { const size_t pos = ch*roi; const size_t drs = ch*1024; DrsCalibrate::ApplyCh(vec+pos, val+pos, start[ch], roi, fOffset.data()+drs, fNumOffset, fGain.data() +drs, fNumGain); } return false; } for (size_t ch=0; ch<1440; ch++) { const size_t pos = ch*fRoi; const size_t drs = ch*1024; DrsCalibrate::ApplyCh(vec+pos, val+pos, start[ch], roi, fOffset.data()+drs, fNumOffset, fGain.data() +drs, fNumGain, fTrgOff.data()+pos, fNumTrgOff); } for (size_t ch=0; ch