void DMRecon::start() { progress.start_time = std::time(0); analyzeFeatures(); globalViewSelection(); processFeatures(); processQueue(); if (progress.cancelled) { progress.status = RECON_CANCELLED; return; } progress.status = RECON_SAVING; SingleViewPtr refV(views[settings.refViewNr]); if (settings.writePlyFile) { refV->saveReconAsPly(settings.plyPath, settings.scale); } refV->writeReconImages(settings.scale); progress.status = RECON_IDLE; // Output percentage of filled pixels { int nrPix = this->width * this->height; float percent = (float) progress.filled / (float) nrPix; std::cout << "Filled " << progress.filled << " pixels, i.e. " << util::string::get_fixed(percent * 100.f, 1) << " %." << std::endl; log << "Filled " << progress.filled << " pixels, i.e. " << util::string::get_fixed(percent * 100.f, 1) << " %." << std::endl; } // Output required time to process the image { size_t mvs_time = std::time(0) - progress.start_time; std::cout << "MVS took " << mvs_time << " seconds." << std::endl; log << "MVS took " << mvs_time << " seconds." << std::endl; } }
void UT_CG729PayloadFormatRead::UT_CG729PayloadFormatRead_CreateSourceBufferL_1L( ) { TBool refO( ETrue ); TBool refV( EFalse ); if ( !iAlloc ) { EUNIT_ASSERT_NO_LEAVE( iRead->CreateSourceBufferL( KUidMediaTypeAudio, *iBuffi, refO ) ); EUNIT_ASSERT_NO_LEAVE( iRead->CreateSourceBufferL( KUidMediaTypeAudio, *iBuffi, refV ) ); EUNIT_ASSERT_SPECIFIC_LEAVE( iRead->CreateSourceBufferL( KUidMediaTypeVideo, *iBuffi, refO ), KErrNotSupported ); } else { EUNIT_ASSERT_NO_LEAVE( iRead->CreateSourceBufferL( KUidMediaTypeAudio, *iBuffi, refO ) ); EUNIT_ASSERT_NO_LEAVE( iRead->CreateSourceBufferL( KUidMediaTypeAudio, *iBuffi, refV ) ); EUNIT_ASSERT_SPECIFIC_LEAVE( iRead->CreateSourceBufferL( KUidMediaTypeVideo, *iBuffi, refO ), KErrNotSupported ); } }