bool TestHypothesesGrow::process() { NCVStatus ncvStat; bool rcode = false; NCVVectorAlloc<Ncv32u> h_vecSrc(*this->allocatorCPU.get(), this->maxLenSrc); ncvAssertReturn(h_vecSrc.isMemAllocated(), false); NCVVectorAlloc<Ncv32u> d_vecSrc(*this->allocatorGPU.get(), this->maxLenSrc); ncvAssertReturn(d_vecSrc.isMemAllocated(), false); NCVVectorAlloc<NcvRect32u> h_vecDst(*this->allocatorCPU.get(), this->maxLenDst); ncvAssertReturn(h_vecDst.isMemAllocated(), false); NCVVectorAlloc<NcvRect32u> d_vecDst(*this->allocatorGPU.get(), this->maxLenDst); ncvAssertReturn(d_vecDst.isMemAllocated(), false); NCVVectorAlloc<NcvRect32u> h_vecDst_d(*this->allocatorCPU.get(), this->maxLenDst); ncvAssertReturn(h_vecDst_d.isMemAllocated(), false); NCV_SET_SKIP_COND(this->allocatorGPU.get()->isCounting()); NCV_SKIP_COND_BEGIN ncvAssertReturn(this->src.fill(h_vecSrc), false); memset(h_vecDst.ptr(), 0, h_vecDst.length() * sizeof(NcvRect32u)); NCVVectorReuse<Ncv32u> h_vecDst_as32u(h_vecDst.getSegment(), lenDst * sizeof(NcvRect32u) / sizeof(Ncv32u)); ncvAssertReturn(h_vecDst_as32u.isMemReused(), false); ncvAssertReturn(this->src.fill(h_vecDst_as32u), false); memcpy(h_vecDst_d.ptr(), h_vecDst.ptr(), h_vecDst.length() * sizeof(NcvRect32u)); NCV_SKIP_COND_END ncvStat = h_vecSrc.copySolid(d_vecSrc, 0); ncvAssertReturn(ncvStat == NCV_SUCCESS, false); ncvStat = h_vecDst.copySolid(d_vecDst, 0); ncvAssertReturn(ncvStat == NCV_SUCCESS, false); ncvAssertCUDAReturn(cudaStreamSynchronize(0), false); Ncv32u h_outElemNum_d = 0; Ncv32u h_outElemNum_h = 0; NCV_SKIP_COND_BEGIN h_outElemNum_d = this->lenDst; ncvStat = ncvGrowDetectionsVector_device(d_vecSrc, this->lenSrc, d_vecDst, h_outElemNum_d, this->maxLenDst, this->rectWidth, this->rectHeight, this->rectScale, 0); ncvAssertReturn(ncvStat == NCV_SUCCESS, false); ncvStat = d_vecDst.copySolid(h_vecDst_d, 0); ncvAssertReturn(ncvStat == NCV_SUCCESS, false); ncvAssertCUDAReturn(cudaStreamSynchronize(0), false); h_outElemNum_h = this->lenDst; ncvStat = ncvGrowDetectionsVector_host(h_vecSrc, this->lenSrc, h_vecDst, h_outElemNum_h, this->maxLenDst, this->rectWidth, this->rectHeight, this->rectScale); ncvAssertReturn(ncvStat == NCV_SUCCESS, false); NCV_SKIP_COND_END //bit-to-bit check bool bLoopVirgin = true; NCV_SKIP_COND_BEGIN if (h_outElemNum_d != h_outElemNum_h) { bLoopVirgin = false; } else { if (memcmp(h_vecDst.ptr(), h_vecDst_d.ptr(), this->maxLenDst * sizeof(NcvRect32u))) { bLoopVirgin = false; } } NCV_SKIP_COND_END if (bLoopVirgin) { rcode = true; } return rcode; }
bool TestCompact::process() { NCVStatus ncvStat; bool rcode = false; NCVVectorAlloc<Ncv32u> h_vecSrc(*this->allocatorCPU.get(), this->length); ncvAssertReturn(h_vecSrc.isMemAllocated(), false); NCVVectorAlloc<Ncv32u> d_vecSrc(*this->allocatorGPU.get(), this->length); ncvAssertReturn(d_vecSrc.isMemAllocated(), false); NCVVectorAlloc<Ncv32u> h_vecDst(*this->allocatorCPU.get(), this->length); ncvAssertReturn(h_vecDst.isMemAllocated(), false); NCVVectorAlloc<Ncv32u> d_vecDst(*this->allocatorGPU.get(), this->length); ncvAssertReturn(d_vecDst.isMemAllocated(), false); NCVVectorAlloc<Ncv32u> h_vecDst_d(*this->allocatorCPU.get(), this->length); ncvAssertReturn(h_vecDst_d.isMemAllocated(), false); NCV_SET_SKIP_COND(this->allocatorGPU.get()->isCounting()); NCV_SKIP_COND_BEGIN ncvAssertReturn(this->src.fill(h_vecSrc), false); for (Ncv32u i=0; i<this->length; i++) { Ncv32u tmp = (h_vecSrc.ptr()[i]) & 0xFF; tmp = tmp * 99 / 255; if (tmp < this->badElemPercentage) { h_vecSrc.ptr()[i] = this->badElem; } } NCV_SKIP_COND_END NCVVectorAlloc<Ncv32u> h_dstLen(*this->allocatorCPU.get(), 1); ncvAssertReturn(h_dstLen.isMemAllocated(), false); Ncv32u bufSize; ncvStat = nppsStCompactGetSize_32u(this->length, &bufSize, this->devProp); ncvAssertReturn(NPPST_SUCCESS == ncvStat, false); NCVVectorAlloc<Ncv8u> d_tmpBuf(*this->allocatorGPU.get(), bufSize); ncvAssertReturn(d_tmpBuf.isMemAllocated(), false); Ncv32u h_outElemNum_h = 0; NCV_SKIP_COND_BEGIN ncvStat = h_vecSrc.copySolid(d_vecSrc, 0); ncvAssertReturn(ncvStat == NPPST_SUCCESS, false); ncvStat = nppsStCompact_32u(d_vecSrc.ptr(), this->length, d_vecDst.ptr(), h_dstLen.ptr(), this->badElem, d_tmpBuf.ptr(), bufSize, this->devProp); ncvAssertReturn(ncvStat == NPPST_SUCCESS, false); ncvStat = d_vecDst.copySolid(h_vecDst_d, 0); ncvAssertReturn(ncvStat == NPPST_SUCCESS, false); ncvStat = nppsStCompact_32u_host(h_vecSrc.ptr(), this->length, h_vecDst.ptr(), &h_outElemNum_h, this->badElem); ncvAssertReturn(ncvStat == NPPST_SUCCESS, false); NCV_SKIP_COND_END //bit-to-bit check bool bLoopVirgin = true; NCV_SKIP_COND_BEGIN if (h_dstLen.ptr()[0] != h_outElemNum_h) { bLoopVirgin = false; } else { for (Ncv32u i=0; bLoopVirgin && i < h_outElemNum_h; i++) { if (h_vecDst.ptr()[i] != h_vecDst_d.ptr()[i]) { bLoopVirgin = false; } } } NCV_SKIP_COND_END if (bLoopVirgin) { rcode = true; } return rcode; }
bool TestHypothesesFilter::process() { NCVStatus ncvStat; bool rcode = false; NCVVectorAlloc<Ncv32u> h_random32u(*this->allocatorCPU.get(), this->numDstRects * sizeof(NcvRect32u) / sizeof(Ncv32u)); ncvAssertReturn(h_random32u.isMemAllocated(), false); Ncv32u srcSlotSize = 2 * this->minNeighbors + 1; NCVVectorAlloc<NcvRect32u> h_vecSrc(*this->allocatorCPU.get(), this->numDstRects*srcSlotSize); ncvAssertReturn(h_vecSrc.isMemAllocated(), false); NCVVectorAlloc<NcvRect32u> h_vecDst_groundTruth(*this->allocatorCPU.get(), this->numDstRects); ncvAssertReturn(h_vecDst_groundTruth.isMemAllocated(), false); NCV_SET_SKIP_COND(this->allocatorCPU.get()->isCounting()); NCV_SKIP_COND_BEGIN ncvAssertReturn(this->src.fill(h_random32u), false); Ncv32u randCnt = 0; Ncv64f randVal; for (Ncv32u i=0; i<this->numDstRects; i++) { h_vecDst_groundTruth.ptr()[i].x = i * this->canvasWidth / this->numDstRects + this->canvasWidth / (this->numDstRects * 4); h_vecDst_groundTruth.ptr()[i].y = i * this->canvasHeight / this->numDstRects + this->canvasHeight / (this->numDstRects * 4); h_vecDst_groundTruth.ptr()[i].width = this->canvasWidth / (this->numDstRects * 2); h_vecDst_groundTruth.ptr()[i].height = this->canvasHeight / (this->numDstRects * 2); Ncv32u numNeighbors = this->minNeighbors + 1 + (Ncv32u)(((1.0 * h_random32u.ptr()[i]) * (this->minNeighbors + 1)) / 0xFFFFFFFF); numNeighbors = (numNeighbors > srcSlotSize) ? srcSlotSize : numNeighbors; //fill in strong hypotheses (2 * ((1.0 * randVal) / 0xFFFFFFFF) - 1) for (Ncv32u j=0; j<numNeighbors; j++) { randVal = (1.0 * h_random32u.ptr()[randCnt++]) / 0xFFFFFFFF; randCnt = randCnt % h_random32u.length(); h_vecSrc.ptr()[srcSlotSize * i + j].x = h_vecDst_groundTruth.ptr()[i].x + (Ncv32s)(h_vecDst_groundTruth.ptr()[i].width * this->eps * (randVal - 0.5)); randVal = (1.0 * h_random32u.ptr()[randCnt++]) / 0xFFFFFFFF; randCnt = randCnt % h_random32u.length(); h_vecSrc.ptr()[srcSlotSize * i + j].y = h_vecDst_groundTruth.ptr()[i].y + (Ncv32s)(h_vecDst_groundTruth.ptr()[i].height * this->eps * (randVal - 0.5)); h_vecSrc.ptr()[srcSlotSize * i + j].width = h_vecDst_groundTruth.ptr()[i].width; h_vecSrc.ptr()[srcSlotSize * i + j].height = h_vecDst_groundTruth.ptr()[i].height; } //generate weak hypotheses (to be removed in processing) for (Ncv32u j=numNeighbors; j<srcSlotSize; j++) { randVal = (1.0 * h_random32u.ptr()[randCnt++]) / 0xFFFFFFFF; randCnt = randCnt % h_random32u.length(); h_vecSrc.ptr()[srcSlotSize * i + j].x = this->canvasWidth + h_vecDst_groundTruth.ptr()[i].x + (Ncv32s)(h_vecDst_groundTruth.ptr()[i].width * this->eps * (randVal - 0.5)); randVal = (1.0 * h_random32u.ptr()[randCnt++]) / 0xFFFFFFFF; randCnt = randCnt % h_random32u.length(); h_vecSrc.ptr()[srcSlotSize * i + j].y = this->canvasHeight + h_vecDst_groundTruth.ptr()[i].y + (Ncv32s)(h_vecDst_groundTruth.ptr()[i].height * this->eps * (randVal - 0.5)); h_vecSrc.ptr()[srcSlotSize * i + j].width = h_vecDst_groundTruth.ptr()[i].width; h_vecSrc.ptr()[srcSlotSize * i + j].height = h_vecDst_groundTruth.ptr()[i].height; } } //shuffle for (Ncv32u i=0; i<this->numDstRects*srcSlotSize-1; i++) { Ncv32u randValLocal = h_random32u.ptr()[randCnt++]; randCnt = randCnt % h_random32u.length(); Ncv32u secondSwap = randValLocal % (this->numDstRects*srcSlotSize-1 - i); NcvRect32u tmp = h_vecSrc.ptr()[i + secondSwap]; h_vecSrc.ptr()[i + secondSwap] = h_vecSrc.ptr()[i]; h_vecSrc.ptr()[i] = tmp; } NCV_SKIP_COND_END Ncv32u numHypothesesSrc = static_cast<Ncv32u>(h_vecSrc.length()); NCV_SKIP_COND_BEGIN ncvStat = ncvGroupRectangles_host(h_vecSrc, numHypothesesSrc, this->minNeighbors, this->eps, NULL); ncvAssertReturn(ncvStat == NCV_SUCCESS, false); NCV_SKIP_COND_END //verification bool bLoopVirgin = true; NCV_SKIP_COND_BEGIN if (numHypothesesSrc != this->numDstRects) { bLoopVirgin = false; } else { std::vector<NcvRect32u> tmpRects(numHypothesesSrc); memcpy(&tmpRects[0], h_vecSrc.ptr(), numHypothesesSrc * sizeof(NcvRect32u)); std::sort(tmpRects.begin(), tmpRects.end()); for (Ncv32u i=0; i<numHypothesesSrc && bLoopVirgin; i++) { if (!compareRects(tmpRects[i], h_vecDst_groundTruth.ptr()[i], this->eps)) { bLoopVirgin = false; } } } NCV_SKIP_COND_END if (bLoopVirgin) { rcode = true; } return rcode; }