int main(int argc, char* argv[]) { // The exit code of the sample application. int exitCode = 0; // Automagically call PylonInitialize and PylonTerminate to ensure the pylon runtime system // is initialized during the lifetime of this object. Pylon::PylonAutoInitTerm autoInitTerm; //cv::Mat theFrame; try { // Create an instant camera object with the camera device found first. //CInstantCamera camera( CTlFactory::GetInstance().CreateFirstDevice()); CBaslerUsbInstantCamera camera( CTlFactory::GetInstance().CreateFirstDevice()); //MJR // Print the model name of the camera. cout << "Using device " << camera.GetDeviceInfo().GetModelName() << endl; // The parameter MaxNumBuffer can be used to control the count of buffers // allocated for grabbing. The default value of this parameter is 10. camera.MaxNumBuffer = 5; // Start the grabbing of c_countOfImagesToGrab images. // The camera device is parameterized with a default configuration which // sets up free-running continuous acquisition. camera.StartGrabbing( c_countOfImagesToGrab); // This smart pointer will receive the grab result data. CGrabResultPtr ptrGrabResult; // Camera.StopGrabbing() is called automatically by the RetrieveResult() method // when c_countOfImagesToGrab images have been retrieved. tStart = clock(); while (camera.IsGrabbing()) { // Wait for an image and then retrieve it. A timeout of 5000 ms is used. camera.RetrieveResult( 5000, ptrGrabResult, TimeoutHandling_ThrowException); // Image grabbed successfully? if (ptrGrabResult->GrabSucceeded()) { // Access the image data. cout << "SizeX: " << ptrGrabResult->GetWidth() << endl; cout << "SizeY: " << ptrGrabResult->GetHeight() << endl; const uint8_t *pImageBuffer = (uint8_t *) ptrGrabResult->GetBuffer(); cout << "Gray value of first pixel: " << (uint32_t) pImageBuffer[0] << endl << endl; #ifdef PYLON_WIN_BUILD // Display the grabbed image. Pylon::DisplayImage(1, ptrGrabResult); #endif // // Save the image using Pylon API // CImagePersistence::Save( ImageFileFormat_Png, "GrabbedImage.png", ptrGrabResult); // suppose your camera is monochrome... get a pointer to pylon image //const pylon::uint8_t *pImageBuffer = (uint8_t *)ptrGrabResult->GetBuffer(); int frameCols = ptrGrabResult->GetWidth(); int frameRows = ptrGrabResult->GetHeight(); // Map the pylon image buffer to a cv::Mat (create a cv::Mat from external buffer) //theFrame = cv::Mat(cv::Size(frameCols, frameRows), CV_8UC1, (void*)pImageBuffer, cv::Mat::AUTO_STEP); //Mat theFrame = cv::Mat(cv::Size(frameCols, frameRows), CV_8UC1, (void*)pImageBuffer); Mat theFrame = cv::Mat(frameRows, frameCols, CV_8UC1); memcpy(theFrame.ptr(),(uint8_t *)ptrGrabResult->GetBuffer(),frameCols*frameRows); // Save openCV Mat frame char str[15]; tCount = clock() - tStart; sprintf(str, "%f seconds", ((float)tCount)/CLOCKS_PER_SEC); printf("%ld: %f seconds\n", tCount, ((float)tCount)/CLOCKS_PER_SEC); sleep(1); tCount = clock() - tStart; sprintf(str, "%f seconds", ((float)tCount)/CLOCKS_PER_SEC); printf("%ld: %f seconds\n", tCount, ((float)tCount)/CLOCKS_PER_SEC); putText(theFrame,str,cvPoint(30,100),FONT_HERSHEY_SIMPLEX,2,cvScalar(200,200,250),3,CV_AA); imwrite("GrabbedImageCV.png",theFrame); // keep a copy of it //cv::Mat myFrame; //theFrame.copyTo(myFrame); // myFrame life cycle is now under your control // // Convert of opencv format and display // CImageFormatConverter fc; // //fc.OutputPixelFormat = PixelType_BGR8packed; // fc.OutputPixelFormat = PixelType_Mono8; // CPylonImage image; // fc.Convert(image, ptrGrabResult); // Mat cv_img = cv::Mat(ptrGrabResult->GetHeight(), ptrGrabResult->GetWidth(), CV_8UC1,(uint8_t*)image.GetBuffer()); // //imshow(src_window,cv_img); // display the image in OpenCV image window // //waitKey(1); } else { cout << "Error: " << ptrGrabResult->GetErrorCode() << " " << ptrGrabResult->GetErrorDescription() << endl; } } } catch (GenICam::GenericException &e) { // Error handling. cerr << "An exception occurred." << endl << e.GetDescription() << endl; exitCode = 1; } // Comment the following two lines to disable waiting on exit. // cerr << endl << "Press Enter to exit." << endl; // while( cin.get() != '\n'); return exitCode; }
void CBaslerFormatConverter::convertMONOtoCVMONO(IplImage *imgDestination,const CGrabResultPtr& imageSource) { imgDestination->nSize = sizeof(IplImage); imgDestination->ID = 0; imgDestination->dataOrder = 0; imgDestination->origin = 0; imgDestination->roi = NULL; imgDestination->maskROI = NULL; imgDestination->imageId = NULL; imgDestination->tileInfo = NULL; memcpy(imgDestination->imageData,imageSource->GetBuffer(),imageSource->GetPayloadSize()); imgDestination->imageDataOrigin = imgDestination->imageData; }
virtual void OnImageGrabbed(CInstantCamera& camera, const CGrabResultPtr& ptrGrabResult) { intptr_t cameraContextValue = ptrGrabResult->GetCameraContext(); CBaslerUsbGrabResultPtr ptrGrabResultUsb = ptrGrabResult; #ifdef PYLON_WIN_BUILD Pylon::DisplayImage(cameraContextValue, ptrGrabResultUsb); #endif if (G_State == Burst) { int _frame_index = 0; _frame_index = _PC_captured_frame_count[cameraContextValue]; cout << "Frame Grabbed #: " << _frame_index << endl; if (IsReadable(ptrGrabResultUsb->ChunkTimestamp)) { _PC_frame_time_table[2 * cameraContextValue + 1][_frame_index] = 0.000000001*(double)ptrGrabResultUsb->ChunkTimestamp.GetValue(); } _Grab_results[cameraContextValue][_frame_index] = ptrGrabResultUsb; _PC_captured_frame_count[cameraContextValue] += 1; } }
void PylonCamera::execute() { if(!mCam) { return; } try { updateParams(); // This smart pointer will receive the grab result data. CGrabResultPtr ptrGrabResult; // Wait for an image and then retrieve it. A timeout of 1000 ms is used. mCam->RetrieveResult(1000, ptrGrabResult); // Image grabbed successfully= if (ptrGrabResult && ptrGrabResult->GrabSucceeded()) { CPylonImage target; // Convert to correct format for vision system CImageFormatConverter converter; converter.OutputPixelFormat = PixelType_BGR8packed; converter.OutputBitAlignment = OutputBitAlignment_MsbAligned; converter.Convert(target, ptrGrabResult); // Now the grab result can be released. The grab result buffer is now // only held by the pylon image. ptrGrabResult.Release(); // Create an RGBImage from converted data RgbImage image(target.GetHeight(), target.GetWidth(), (uchar *)target.GetBuffer()); mOut.send(image.clone()); // The Release() method can be used to release any data. target.Release(); } } catch (GenICam::GenericException &e) { /*stop(); deinitialize();*/ throw std::runtime_error(e.GetDescription()); } catch(const std::exception& e) { throw std::runtime_error(e.what()); } }
void CCameraBasler::continuousAcquireFreeRunning(int index) { IplImage *img; CGrabResultPtr ptrGrabResult; std::pair<char*,IplImage*> pairTemp; unsigned int indiceImagenes = 0; while (m_cameraState == CVCCameraInterface::GRABBING) { try { m_Camera->RetrieveResult(INFINITE,ptrGrabResult,TimeoutHandling_Return); if (ptrGrabResult->GrabSucceeded()) { img = cvCreateImage(cvSize(ptrGrabResult->GetWidth(),ptrGrabResult->GetHeight()),8,3); CBaslerFormatConverter::convertCOLORtoCVRGB(img,ptrGrabResult); // Construct image file name char *str = new char[100]; sprintf(str,"Images\\V\\Visible%d.png",indiceImagenes); // Put name and image in a pair pairTemp.first = str; pairTemp.second = img; // Enqueue pair of <nameOfFileToSave,image> in the SaveImageProcess // Is a singleton thread save design to enqueue and process all images. // Index refers to the "id" of the queue. Each queue will have a thread. CVCCameraInterface::SaveImagesProcess::Instance()->Add(index,pairTemp); // Increment the number of the actual image. Used to make the filename indiceImagenes++; } else { std::cout << "Error: [Basler] :" << ptrGrabResult->GetErrorCode() << " " << ptrGrabResult->GetErrorDescription() << std::endl; } } catch(...) { std::cerr << "ERROR: [Basler] Acquiring continuous images." << std::endl; } } try { m_Camera->StopGrabbing(); } catch(...) { std::cerr << "ERROR: [Basler] Stopping continuous acquisition." << std::endl; } }
int main(int argc, char* argv[]) { // The exit code of the sample application. int exitCode = 0; // Automagically call PylonInitialize and PylonTerminate to ensure the pylon runtime system // is initialized during the lifetime of this object. Pylon::PylonAutoInitTerm autoInitTerm; try { // Create an instant camera object with the camera device found first. CInstantCamera camera( CTlFactory::GetInstance().CreateFirstDevice()); // Print the model name of the camera. cout << "Using device " << camera.GetDeviceInfo().GetModelName() << endl; // The parameter MaxNumBuffer can be used to control the count of buffers // allocated for grabbing. The default value of this parameter is 10. camera.MaxNumBuffer = 5; // Start the grabbing of c_countOfImagesToGrab images. // The camera device is parameterized with a default configuration which // sets up free-running continuous acquisition. camera.StartGrabbing( c_countOfImagesToGrab); // This smart pointer will receive the grab result data. CGrabResultPtr ptrGrabResult; // Camera.StopGrabbing() is called automatically by the RetrieveResult() method // when c_countOfImagesToGrab images have been retrieved. while ( camera.IsGrabbing()) { // Wait for an image and then retrieve it. A timeout of 5000 ms is used. camera.RetrieveResult( 5000, ptrGrabResult, TimeoutHandling_ThrowException); // Image grabbed successfully? if (ptrGrabResult->GrabSucceeded()) { // Access the image data. cout << "SizeX: " << ptrGrabResult->GetWidth() << endl; cout << "SizeY: " << ptrGrabResult->GetHeight() << endl; const uint8_t *pImageBuffer = (uint8_t *) ptrGrabResult->GetBuffer(); cout << "Gray value of first pixel: " << (uint32_t) pImageBuffer[0] << endl << endl; #ifdef PYLON_WIN_BUILD // Display the grabbed image. Pylon::DisplayImage(1, ptrGrabResult); #endif } else { cout << "Error: " << ptrGrabResult->GetErrorCode() << " " << ptrGrabResult->GetErrorDescription() << endl; } } } catch (GenICam::GenericException &e) { // Error handling. cerr << "An exception occurred." << endl << e.GetDescription() << endl; exitCode = 1; } // Comment the following two lines to disable waiting on exit. cerr << endl << "Press Enter to exit." << endl; while( cin.get() != '\n'); return exitCode; }
int main(int argc, char* argv[]) { ros::init(argc, argv, "grab"); ros::NodeHandle n; ros::Rate loop_rate(10); // The exit code of the sample application. int exitCode = 0; // Automagically call PylonInitialize and PylonTerminate to ensure // the pylon runtime system is initialized during the lifetime of this object. Pylon::PylonAutoInitTerm autoInitTerm; CGrabResultPtr ptrGrabResult; namedWindow("CV_Image",WINDOW_AUTOSIZE); try { CInstantCamera camera( CTlFactory::GetInstance().CreateFirstDevice()); cout << "Using device " << camera.GetDeviceInfo().GetModelName() << endl; camera.Open(); GenApi::CIntegerPtr width(camera.GetNodeMap().GetNode("Width")); GenApi::CIntegerPtr height(camera.GetNodeMap().GetNode("Height")); Mat cv_img(width->GetValue(), height->GetValue(), CV_8UC3); camera.StartGrabbing(); CPylonImage image; CImageFormatConverter fc; fc.OutputPixelFormat = PixelType_RGB8packed; while(camera.IsGrabbing()){ camera.RetrieveResult( 500, ptrGrabResult, TimeoutHandling_ThrowException); if (ptrGrabResult->GrabSucceeded()){ fc.Convert(image, ptrGrabResult); cv_img = cv::Mat(ptrGrabResult->GetHeight(), ptrGrabResult->GetWidth(), CV_8UC3,(uint8_t*)image.GetBuffer()); imshow("CV_Image",cv_img); waitKey(1); if(waitKey(30)==27){ camera.StopGrabbing(); } } ros::spinOnce(); loop_rate.sleep(); } } catch (GenICam::GenericException &e) { // Error handling. cerr << "An exception occurred." << endl << e.GetDescription() << endl; exitCode = 1; } // Comment the following two lines to disable waiting on exit cerr << endl << "Press Enter to exit." << endl; while( cin.get() != '\n'); return exitCode; }
int main(int argc, char* argv[]) { //------------------------// // Main Code // //------------------------// // The exit code of the sample application. int exitCode = 0; system("CLS"); cout << "*------------------------------------------------------*" << endl; cout << "******* Program by S.E Lansbergen, June 2016 ********" << endl; cout << "*------------------------------------------------------*" << endl; //------------------------// // Configuration variable // //------------------------// // Config file const char* open_config = "config_grabpupilsize.cfg"; // INI file (containing default start-up values) const char* open_ini = "default.ini"; //------------------------// // R/W default ini file // //------------------------// string def_prethres; string def_mainthres; string def_roiheigth; string def_roiwidth; string def_roistartx; string def_roistarty; string def_itterations; string def_pupil_max; string def_pupil_min; // create or open an ini file INI ini(open_ini); // load INI file if (ini.is_open()) { // The file exists, and is open for input // get default threshold values def_prethres = ini.property("thresholding", "pre-threshold"); def_mainthres = ini.property("thresholding", "main-threshold"); // get default ROI values def_roiheigth = ini.property("ROI", "height"); def_roiwidth = ini.property("ROI", "width"); def_roistartx = ini.property("ROI", "startx"); def_roistarty = ini.property("ROI", "starty"); // get default closing itteration value def_itterations = ini.property("close", "itterations"); // get default min and max pupil size values def_pupil_max = ini.property("pupil_size", "pupil_max"); def_pupil_min = ini.property("pupil_size", "pupil_min"); cout << endl << endl << " *** default.ini loaded ***" << endl; } else { cout << endl << " *** default.ini file is missing, program aborted! *** " << endl; cout << " *** go to Github and download default.ini, put this in the executable root dir *** " << endl; return 1; } //-------------------// // Defualt variables // //-------------------// // default hard coded settings if Config.cfg file is not present or in-complete/commented const char* thres_name = "THRESH_BINARY"; const char* save_path = "c:/output.avi"; const char* save_path_num = "c:/output.txt"; const char* save_path_ori = "c:/original.avi"; string sav_pat = "c:/output.avi"; string sav_pat_num = "c:/output.txt";; string sav_pat_ori = "c:/original.avi"; string save_path_xy = "c:/xy_postion.txt"; bool show_screen_info = 1; bool save_video = 0; bool save_radius = 0; bool save_original = 0; bool ROI_start_auto = 1; bool original_image = 1; bool blurred_image = 0; bool thresholded_image = 0; bool closed_image = 0; bool end_result_image = 1; bool show_ost = 0; bool pre_threshold_image = 0; bool crosshair = 0; bool calibrate = 0; ThresholdTypes thres_type = THRESH_BINARY; int cFramesPerSecond = 20; int main_pre_threshold = 20; int thres_mode = 0; int ROI_start_x; int ROI_start_y; int width_SE = 3; int heigth_SE = 3; int heigth_blur = 3; int width_blur = 3; int pupil_min = 15; int pupil_max = 55; int pre_threshold = 100; int itterations_close = 3; int cali_x_a = 280; int cali_x_b = 350; int cali_y_a = 100; int cali_y_b = 200; Size ROI_dimensions = Size(150, 150); Size ROI_start; Size blur_dimensions = Size(width_SE, heigth_SE); Size SE_morph = Size(width_SE, heigth_SE); double pupil_aspect_ratio = 1.5; double size_text = 0.5; double pi = 3.14159; uint32_t time, frames; Point pupil_position; //---------------------// // Remaining variables // //---------------------// // video writer output VideoWriter roi_end_result, original; // images for processing Mat thres; Mat close; Mat blur; Mat roi_eye; Mat pre_thres; Mat aim; Mat cali; // variables for numerical output ofstream output_end_result, output_xy; ostringstream strs, ost1, ost2, ost3; string radius, size_roi, frame_rate, output_file; //-------------------------// // read configuration file // //-------------------------// // read config file ifstream ifile(open_config); if (ifile) { // The file exists, and is open for input ConfigFile cfg(open_config); // check for existance and replace defualt value if (cfg.keyExists("show_screen_info") == true) { // get screen info on/off show_screen_info = cfg.getValueOfKey<bool>("show_screen_info"); } if (cfg.keyExists("save_video") == true) { // get video save_file info save_video = cfg.getValueOfKey<bool>("save_video"); } if (cfg.keyExists("save_radius") == true) { // get numerical save_file info save_radius = cfg.getValueOfKey<bool>("save_radius"); } if (cfg.keyExists("save_original") == true) { // get save original stream info save_original = cfg.getValueOfKey<bool>("save_original"); } if (cfg.keyExists("frames_per_sec") == true) { // get frames per second cFramesPerSecond = cfg.getValueOfKey<int>("frames_per_sec"); } if (cfg.keyExists("save_path_vid") == true) { // get video output file name & path sav_pat = cfg.getValueOfKey<string>("save_path_vid"); save_path = sav_pat.c_str(); } if (cfg.keyExists("save_path_num") == true) { // get numerical output file name & path sav_pat_num = cfg.getValueOfKey<string>("save_path_num"); save_path_num = sav_pat_num.c_str(); } if (cfg.keyExists("save_path_ori") == true) { // get original stream file name & path sav_pat_ori = cfg.getValueOfKey<string>("save_path_ori"); save_path_ori = sav_pat_ori.c_str(); } if (cfg.keyExists("height_roi") == true && cfg.keyExists("width_roi") == true) { // get heigth & width ROI int ROI_heigth = cfg.getValueOfKey<int>("height_roi"); int ROI_width = cfg.getValueOfKey<int>("width_roi"); ROI_dimensions = Size(ROI_width, ROI_heigth); } if (cfg.keyExists("ROI_start_x") == true && cfg.keyExists("ROI_start_y") == true) { // get x and y starting point for ROI ROI_start_y = cfg.getValueOfKey<int>("ROI_start_y"); ROI_start_x = cfg.getValueOfKey<int>("ROI_start_x"); ROI_start = Size(ROI_start_x, ROI_start_y); ROI_start_auto = 0; } if (cfg.keyExists("width_SE") == true && cfg.keyExists("heigth_SE") == true) { // get dimensions SE width_SE = cfg.getValueOfKey<int>("width_SE"); heigth_SE = cfg.getValueOfKey<int>("heigth_SE"); SE_morph = Size(width_SE, heigth_SE); } if (cfg.keyExists("heigth_blur") == true && cfg.keyExists("width_blur") == true) { // get dimensions Gaussian blur heigth_blur = cfg.getValueOfKey<int>("heigth_blur"); width_blur = cfg.getValueOfKey<int>("width_blur"); blur_dimensions = Size(width_blur, heigth_blur); } if (cfg.keyExists("thres_mode") == true) { // get threshold method thres_mode = cfg.getValueOfKey<int>("thres_mode"); switch (thres_mode) { case 0: thres_type = THRESH_BINARY; thres_name = "THRESH_BINARY"; break; case 1: thres_type = THRESH_BINARY_INV; thres_name = "THRESH_BINARY_INV"; break; case 2: thres_type = THRESH_TRUNC; thres_name = "THRESH_TRUNC"; break; case 3: thres_type = THRESH_TOZERO; thres_name = "THRESH_TOZERO"; break; case 4: thres_type = THRESH_TOZERO_INV; thres_name = "THRESH_TOZERO_INV"; break; default: thres_type = THRESH_BINARY; thres_name = "THRESH_BINARY"; } } if (cfg.keyExists("itterations_close") == true) { // get number of itterations for closing operation itterations_close = cfg.getValueOfKey<int>("itterations_close"); } if (cfg.keyExists("pupil_aspect_ratio") == true) { // get aspect ratio threshold accepted ellipse pupil_aspect_ratio = cfg.getValueOfKey<double>("pupil_aspect_ratio"); } if (cfg.keyExists("pupil_min") == true) { // get minimal accepted pupil radius pupil_min = cfg.getValueOfKey<int>("pupil_min"); } if (cfg.keyExists("pupil_max") == true) { // get maximal accepted pupil radius pupil_max = cfg.getValueOfKey<int>("pupil_max"); } if (cfg.keyExists("original_image") == true) { // info: stream original stream to display original_image = cfg.getValueOfKey<bool>("original_image"); } if (cfg.keyExists("blurred_image") == true) { // info: stream blurred stream to display blurred_image = cfg.getValueOfKey<bool>("blurred_image"); } if (cfg.keyExists("thresholded_image") == true) { // info: thresholded_image blurred stream to display thresholded_image = cfg.getValueOfKey<bool>("thresholded_image"); } if (cfg.keyExists("closed_image") == true) { // info: closed_image blurred stream to display closed_image = cfg.getValueOfKey<bool>("closed_image"); } if (cfg.keyExists("end_result_image") == true) { // info: end_result_image blurred stream to display end_result_image = cfg.getValueOfKey<bool>("end_result_image"); } if (cfg.keyExists("show_ost") == true) { // put text on screen info show_ost = cfg.getValueOfKey<bool>("show_ost"); } if (cfg.keyExists("size_text") == true) { // get text size for on screen text size_text = cfg.getValueOfKey<double>("size_text"); } if (cfg.keyExists("threshold") == true) { // get threshold value main_pre_threshold = cfg.getValueOfKey<int>("threshold"); } // BETA 31-5-2016, 3-6-2016 if (cfg.keyExists("pre_threshold") == true) { // get pre threshold value pre_threshold = cfg.getValueOfKey<int>("pre_threshold"); } if (cfg.keyExists("pre_threshold_image") == true) { // info: pre_threshold_image stream to display pre_threshold_image = cfg.getValueOfKey<bool>("pre_threshold_image"); } // Beta 31-5-2016 if (cfg.keyExists("crosshair") == true) { // crosshair = cfg.getValueOfKey<bool>("crosshair"); } // Beta 1-6-2016 if (cfg.keyExists("calibrate") == true) { // calibrate = cfg.getValueOfKey<bool>("calibrate"); } if (cfg.keyExists("cali_x_a") == true && cfg.keyExists("cali_x_b") == true && cfg.keyExists("cali_y_a") == true && cfg.keyExists("cali_y_b") == true) { // cali_x_a = cfg.getValueOfKey<int>("cali_x_a"); cali_x_b = cfg.getValueOfKey<int>("cali_x_b"); cali_y_a = cfg.getValueOfKey<int>("cali_y_a"); cali_y_b = cfg.getValueOfKey<int>("cali_y_b"); } // BETA 14-6-2016 if (cfg.keyExists("save_path_xy") == true) { // get numerical output file name & path save_path_xy = cfg.getValueOfKey<string>("save_path_xy"); //save_path_num = sav_pat_num.c_str(); } cout << endl << endl << " *** Configuration file loaded ***" << endl; } else { cout << endl << endl << " *** No configuration file found ***" << endl; cout << " *** Default, internal parameters loaded ***" << endl; } //------------------------// // Handle input arguments // //------------------------// cout << endl << endl << "Program has " << (argc - 1) << " extra input arguments:" << endl; for (int i = 1; i < argc; ++i) cout << argv[i] << endl; //------------------------// // CMD - parser // //------------------------// // input: none if (argc <= 1) // actual count (starts at 1) { // disable saving both numerical as well as video output. save_video = false; //save_video = false; save_radius = false; // amount of frames to grab frames = 120 * cFramesPerSecond; cout << endl << endl << " *** Calibration mode *** " << endl << endl; cout << " - No recording of video or numerical output" << endl; cout << " - 120 seconds of video stream" << endl << endl; cout << " - Hit [ESC] in video window to quit the program" << endl; } // input: [time] else if (argc == 2) { if (isalpha(*argv[1])) { // index starts at 0 puts("Not a number: Wrong input argument for [time] ?"); } else if (isdigit(*argv[1])) { // atoi = Ascii To Int -> stops converting at the first non int time = atoi(argv[1]); // amount of frames to grab frames = time * cFramesPerSecond; cout << endl << "Recording time (sec(s)) : " << time << endl; cout << "Total frames recorded : " << frames << endl; cout << "No additional savepath and name entered for numeric output" << endl; cout << "No additional savepath and name entered for video output." << endl; } else { puts("Not a number: Wrong input argument for [time] ?"); return 1; } } // input: [time] [save num] else if (argc == 3) { if (isalpha(*argv[1]) || isdigit(*argv[2])) { // index starts at 0 puts("Not a number: Wrong input argument for [time] or [save num] ?"); } else if (isdigit(*argv[1]) && isalpha(*argv[2])) { // atoi = Ascii To Int -> stops converting at the first non int time = atoi(argv[1]); // amount of frames to grab frames = time * cFramesPerSecond; // get entered save path and name for numerical output save_path_num = argv[2]; cout << endl << "Recording time (sec(s)) : " << time << endl; cout << "Total frames recorded : " << frames << endl; cout << "Entered additional savepath and name for video output : " << save_path_num << endl; cout << "No additional savepath and name entered for video output." << endl; } else { puts("Not a number: Wrong input argument for [time] or [save num] ?"); return 1; } } // input: [time] [save num] [save xy] else if (argc == 4) { // atoi = Ascii To Int -> stops converting at the first non int time = atoi(argv[1]); // amount of frames to grab frames = time * cFramesPerSecond; // get entered save path and name for pupil area output save_path_num = argv[2]; // get entered save path and name for xy position pupil output save_path_xy = argv[3]; cout << endl << "Recording time (sec(s)) : " << time << endl; cout << "Total frames recorded : " << frames << endl; cout << "Entered additional savepath and name for pupil area output : " << save_path_num << endl; cout << "Entered additional savepath and name for xy position pupil : " << save_path_xy << endl; } // input: [time] [save num] [save xy] [save vid] else if (argc == 5) { // atoi = Ascii To Int -> stops converting at the first non int time = atoi(argv[1]); // amount of frames to grab frames = time * cFramesPerSecond; // get entered save path and name for numerical output save_path_num = argv[2]; // get entered save path and name for xy position pupil output save_path_xy = argv[3]; // get entered save path and name for video output save_path = argv[4]; cout << endl << "Recording time (sec(s)) : " << time << endl; cout << "Total frames recorded : " << frames << endl; cout << "Entered additional savepath and name for pupil area output : " << save_path_num << endl; cout << "Entered additional savepath and name for xy position pupil : " << save_path_xy << endl; cout << "Entered additional savepath and name for video output : " << save_path << endl; } // to many input arguments else { cout << endl << " *** To many input arguments *** " << endl; return 1; } //------------------------------------// // Read values from default.ini file // //------------------------------------// if (ini.is_open()) { main_pre_threshold = atoi(def_mainthres.c_str()); pre_threshold = atoi(def_prethres.c_str()); itterations_close = atoi(def_itterations.c_str()); pupil_min = atoi(def_pupil_min.c_str()); pupil_max = atoi(def_pupil_max.c_str()); ROI_dimensions.width = atoi(def_roiwidth.c_str()); ROI_dimensions.height = atoi(def_roiheigth.c_str()); } //-----------------------// // Show loaded variables // //-----------------------// if (show_screen_info == true) { cout << endl << endl; cout << "*------------------------------------------------------*" << endl; cout << "***** Program Parameters *****" << endl; cout << "*------------------------------------------------------*" << endl; cout << "*" << endl; cout << "* Show Crosshair to aim camera : " << crosshair << endl; cout << "* Save video output : " << save_video << endl; cout << "* Save original stream : " << save_original << endl; cout << "* Save radius numerical output : " << save_radius << endl; cout << "* Save path video output : " << save_path << endl; cout << "* Save path original stream : " << save_path_ori << endl; cout << "* Save path pupil area output : " << save_path_num << endl; cout << "* Save path xy position output : " << save_path_xy << endl; cout << endl; cout << "* Frames per second : " << cFramesPerSecond << endl; cout << "* Heigth and width ROI : " << ROI_dimensions << endl; if (ROI_start_auto == false) { cout << "* Anchor coordinate [X,Y] ROI \n manually set : " << ROI_start << endl; } else { cout << "* Anchor coordinate [X,Y] ROI set automatically" << endl; } cout << endl; cout << "* Value of threshold : " << main_pre_threshold << endl; cout << "* Threshold mode : (" << thres_mode << ") " << thres_name << endl; cout << "* pre-threshold value : " << pre_threshold << endl; cout << "* Size Gaussian blur filter : " << blur_dimensions << endl; cout << "* Size structuring element \n for morphological closing : " << SE_morph << endl; cout << "* Total itterations closing operation : " << itterations_close << endl; cout << "* Threshold aspect ratio ellipse : " << pupil_aspect_ratio << endl; cout << "* Minimum radius accepted ellipse : " << pupil_min << endl; cout << "* Maximum radius accepted ellipse : " << pupil_max << endl; cout << endl; cout << "* Show original stream on display : " << original_image << endl; cout << "* Show blurred stream on display : " << blurred_image << endl; cout << "* Show pre-threshold stream on display: " << pre_threshold_image << endl; cout << "* Show thresholded stream on display : " << thresholded_image << endl; cout << "* Show morph closed stream on display : " << closed_image << endl; cout << "* Show end result stream on display : " << end_result_image << endl; cout << endl; cout << "* Show text on end result stream : " << show_ost << endl; cout << "* Size text on screen : " << size_text << endl; cout << "*" << endl; cout << "*------------------------------------------------------*" << endl; cout << "******* Program by S.E Lansbergen, July 2016 ******** " << endl; cout << "*------------------------------------------------------*" << endl; } // set variables for on screen text output if (show_ost == true) { // text with save path & name ost3 << save_path; output_file = ost3.str(); } if (crosshair == true) { // text for ROI size ost1 << "ROI size: " << ROI_dimensions.width << "x" << ROI_dimensions.height; size_roi = ost1.str(); } //--------------------------------// // Video acquisition and analysis // //--------------------------------// // Before using any pylon methods, the pylon runtime must be initialized. PylonInitialize(); try { // Create an instant camera object with the camera device found first. //CInstantCamera camera(CTlFactory::GetInstance().CreateFirstDevice()); CBaslerUsbInstantCamera camera(CTlFactory::GetInstance().CreateFirstDevice()); // Print the model name of the camera. cout << endl << endl << "Connected Basler USB 3.0 device, type : " << camera.GetDeviceInfo().GetModelName() << endl; // open camera object to parse frame# etc. camera.Open(); // Enable the acquisition frame rate parameter and set the frame rate. camera.AcquisitionFrameRateEnable.SetValue(true); camera.AcquisitionFrameRate.SetValue(cFramesPerSecond); // Get native width and height from connected camera GenApi::CIntegerPtr width(camera.GetNodeMap().GetNode("Width")); GenApi::CIntegerPtr height(camera.GetNodeMap().GetNode("Height")); // The parameter MaxNumBuffer can be used to control the count of buffers // allocated for grabbing. The default value of this parameter is 10. camera.MaxNumBuffer = 5; // Start the grabbing of c_countOfImagesToGrab images. // The camera device is parameterized with a default configuration which // sets up free-running continuous acquisition. camera.StartGrabbing(frames); // convers pylon video stream into CPylonImage object // which will be used to construct Mat objects CPylonImage image; CImageFormatConverter fc; // define pixel output format (to match algorithm optimalization). fc.OutputPixelFormat = PixelType_Mono8; // This smart pointer will receive the grab result data. (Pylon). CGrabResultPtr ptrGrabResult; // create Mat image template Mat cv_img(width->GetValue(), height->GetValue(), CV_8UC3); // set contours variable vector<vector<Point> > contours; // automatically calculate start point of ROI if (ROI_start_auto == true) { Size size(width->GetValue(), height->GetValue()); ROI_start_x = (size.width / 2) - (ROI_dimensions.width / 2); ROI_start_y = (size.height / 2) - (ROI_dimensions.height / 2); // get values from default.ini if (ini.is_open()) { ROI_start_x = atoi(def_roistartx.c_str()); ROI_start_y = atoi(def_roistarty.c_str()); } } // set ROI Rect roi(ROI_start_x, ROI_start_y, ROI_dimensions.width, ROI_dimensions.height); // set Structuring Element Mat SE = getStructuringElement(MORPH_ELLIPSE, SE_morph, Point(-1, -1)); //--------------------------------// // Open avi and txt output stream // //--------------------------------// // open video writer object: End Result if (save_video == true) { // create video output End Result object -> MPEG encoding roi_end_result.open(save_path, CV_FOURCC('M', 'P', 'E', 'G'), cFramesPerSecond, Size(width->GetValue(), height->GetValue()), true); //if the VideoWriter file is not initialized successfully, exit the program. if (!roi_end_result.isOpened()) { cout << "ERROR: Failed to write the video" << endl; return -1; } } // open video writer object: Original if (save_original == true) { // create video output Original object -> MPEG encoding original.open(save_path_ori, CV_FOURCC('M', 'P', 'E', 'G'), cFramesPerSecond, Size(width->GetValue(), height->GetValue()), true); //if the VideoWriter file is not initialized successfully, exit the program. if (!original.isOpened()) { cout << "ERROR: Failed to write the video" << endl; return -1; } } // open outstream to write end result (radius) if (save_radius == true) { output_end_result.open(save_path_num); output_xy.open(save_path_xy); } // set parameters for calibration tool Size size_cam(width->GetValue(), height->GetValue()); Point calibrate_stripe1a(cali_x_a, cali_y_a); Point calibrate_stripe1b(cali_x_a, cali_y_b); Point calibrate_stripe2a(cali_x_b, cali_y_a); Point calibrate_stripe2b(cali_x_b, cali_y_b); //---------------------// // Main Algorithm Loop // //---------------------// // initiate main loop with algorithm while (camera.IsGrabbing()) { // Wait for an image and then retrieve it. A timeout of 5000 ms is used. camera.RetrieveResult(5000, ptrGrabResult, TimeoutHandling_ThrowException); // Image grabbed successfully? if (ptrGrabResult->GrabSucceeded()) { // Pre-Step: set (click on-)mouse call back function setMouseCallback("End Result", onMouse, 0); // Step 1 // // convert to Mat - openCV format for analysis fc.Convert(image, ptrGrabResult); cv_img = Mat(ptrGrabResult->GetHeight(), ptrGrabResult->GetWidth(), CV_8U, (uint8_t*)image.GetBuffer()); // (Step 1b) // // crosshair output if (crosshair == true) { cvtColor(cv_img, aim, CV_GRAY2RGB); line(aim, Point((size_cam.width / 2 - 25), size_cam.height / 2), Point((size_cam.width / 2 + 25), size_cam.height / 2), Scalar(0, 255, 0), 2, 8); line(aim, Point(size_cam.width / 2, (size_cam.height / 2 - 25)), Point(size_cam.width / 2, (size_cam.height / 2 + 25)), Scalar(0, 255, 0), 2, 8); rectangle(aim, roi, Scalar(0, 255, 0), 2, 8); putText(aim, size_roi, cvPoint(30, 20), FONT_HERSHEY_COMPLEX_SMALL, (0.8), cvScalar(0, 255, 0), 1, CV_AA); imshow("Crosshair", aim); } // (Step 1c) // // extra calibration step output if (calibrate == true) { cvtColor(cv_img, cali, CV_GRAY2RGB); line(cali, calibrate_stripe1a, calibrate_stripe1b, Scalar(255, 0, 0), 4, 8); line(cali, calibrate_stripe2a, calibrate_stripe2b, Scalar(255, 0, 0), 4, 8); putText(cali, "Calibrate", cvPoint(30, 20), FONT_HERSHEY_COMPLEX_SMALL, (0.8), cvScalar(255, 0, 0), 1, CV_AA); imshow("calibrate", cali); } // Step 2 // // take ROI from eye original eye = cv_img; roi_eye = eye(roi); // make RGB copy of ROI for end result cvtColor(eye, end_result, CV_GRAY2RGB); // Set user ROI from mouse input if (roi_user.width != 0) { if (0 <= roi_user.x && 0 <= roi_user.width && roi_user.x + roi_user.width <= eye.cols && roi_user.y + roi_user.height <= eye.rows) { roi = roi_user; } } if (original_image == true) { imshow("Original", eye); } // Step 3 // // apply Gaussian blur to ROI GaussianBlur(roi_eye, blur, blur_dimensions, 0, 0, BORDER_DEFAULT); if (blurred_image == true) { imshow("Gaussian blur", blur); } // Step 4 // // Pre-Threshold: Convert to binary image by thresholding it threshold(blur, pre_thres, pre_threshold, 255, THRESH_TOZERO_INV); if (pre_threshold_image == true) { imshow("Pre-thresholded", pre_thres); // set trackbar on end-result output createTrackbar("Pre Threshold:", "End Result", &pre_threshold, 255); } // Step 5 // // Main-Threshold: Convert to binary image by thresholding it threshold(pre_thres, thres, main_pre_threshold, 255, thres_type); if (thresholded_image == true) { imshow("Thresholded", thres); // set trackbar on end-result output createTrackbar(" Threshold:", "End Result", &main_pre_threshold, 255); } // Step 6 // // Morphological closing (erosion and dilation) morphologyEx(thres, close, MORPH_CLOSE, SE, Point(-1, -1), itterations_close); if (closed_image == true) { imshow("Morphological closing", close); // set trackbar on end-result output createTrackbar(" Itterations:", "End Result", &itterations_close, 15); } // Step 7 // // find contour algorithm findContours(close, contours, RETR_LIST, CHAIN_APPROX_NONE); // Step 8 // // Fit ellipse and draw on image double ellipse_width(0), ellipse_height(0); int flag(0), area_output; Point ellipse_shift(0, 0); int x_out(0), y_out(0); // Loop to draw circle on video image for (int i = 0; i < contours.size(); i++) { size_t count = contours[i].size(); if (count < 6) continue; Mat pointsf; Mat(contours[i]).convertTo(pointsf, CV_32F); RotatedRect box = fitEllipse(pointsf); if (MAX(box.size.width, box.size.height) > MIN(box.size.width, box.size.height) * pupil_aspect_ratio) continue; // sets min and max width and heigth of the box in which the ellipse is fitted // only these are used in the video and numerical output if (MAX(box.size.width, box.size.height) > pupil_min && MAX(box.size.width, box.size.height) < pupil_max) { flag++; // counts 1 at first itteration // adds all width and height in all itterations for pupil area ellipse_width = ellipse_width + box.size.width; ellipse_height = ellipse_height + box.size.height; // Plot on ROI screen, and add shift x-y by ROI ellipse_shift.x = box.center.x + roi.x; ellipse_shift.y = box.center.y + roi.y; ellipse(end_result, ellipse_shift, box.size*0.5f, box.angle, 0, 360, Scalar(0, 0, 255), 2, LINE_AA); // adds all width and height in all itterations for pupil x-y position x_out = x_out + ellipse_shift.x; y_out = y_out + ellipse_shift.y; } } // draw cross in center of pupil line(end_result, Point(ellipse_shift.x - 3, ellipse_shift.y), Point(ellipse_shift.x + 3, ellipse_shift.y), Scalar(0, 255, 0), 2, 8); line(end_result, Point(ellipse_shift.x, ellipse_shift.y - 3), Point(ellipse_shift.x, ellipse_shift.y + 3), Scalar(0, 255, 0), 2, 8); // devides width and heigth with total number of found ellipses to get average value // if (ellipse_width != NAN && ellipse_height != NAN && ellipse_width != 0 && ellipse_height != 0) { ellipse_width = ellipse_width / flag; ellipse_height = ellipse_height / flag; // calculate total area of ellipse area_output = (ellipse_width / 2) * (ellipse_height / 2) * pi; // set x-y position x_out = x_out / flag; y_out = y_out / flag; } else { // set area to 0 when no ellipse is found area_output = 0; x_out = 0; y_out = 0; } // put streamstring to video frame if (show_ost == true) { putText(end_result, output_file, cvPoint(30, 20), FONT_HERSHEY_COMPLEX_SMALL, (size_text), cvScalar(0, 0, 255), 1, CV_AA); } // SHOW END RESULT // // set trackbar on end-result output imshow("End Result", end_result); createTrackbar("size min:", "End Result", &pupil_min, 75); createTrackbar("size max:", "End Result", &pupil_max, 150); //------------------------------// // Store radius & video streams // //------------------------------// // store radius in output file if (save_radius == true) { output_end_result << area_output << endl; output_xy << x_out << char(44) << char(32) << y_out << endl; } // write the end result into file if (save_video == true) { roi_end_result.write(end_result); } // write the original stream into file if (save_original == true) { // conversion of Mat file is necassery prior to saving with mpeg compression cvtColor(eye, eye, CV_GRAY2RGB); original.write(eye); } // close grabbing when escape-key "Esc" is used if (waitKey(30) == 27) { camera.StopGrabbing(); } } else { cout << "Error: " << ptrGrabResult->GetErrorCode() << " " << ptrGrabResult->GetErrorDescription() << endl; } } // end algorithm main loop // close numerical output streams if (save_radius == true) { output_end_result.close(); output_xy.close(); } //------------------------------------// // Write values from default.ini file // //------------------------------------// if (ini.is_open()) { // set default threshold values ini.property("thresholding", "pre-threshold") = to_string(pre_threshold); ini.property("thresholding", "main-threshold") = to_string(main_pre_threshold); // set default ROI values ini.property("ROI", "width") = to_string(roi.width); ini.property("ROI", "height") = to_string(roi.height); ini.property("ROI", "startx") = to_string(roi.x); ini.property("ROI", "starty") = to_string(roi.y); // get default closing itteration value ini.property("close", "itterations") = to_string(itterations_close); // get default min and max pupil size values ini.property("pupil_size", "pupil_max") = to_string(pupil_max); ini.property("pupil_size", "pupil_min") = to_string(pupil_min); cout << endl << endl << " *** default.ini stored ***" << endl; } } catch (const GenericException &e) { // Error handling. cerr << "An exception occurred." << endl << e.GetDescription() << endl; exitCode = 1; } // Releases all pylon resources. PylonTerminate(); // end on terminal screen if (save_radius == true || save_video == true) { cout << endl << endl << " *** Done recording ***" << endl << endl; } else { cout << endl << endl << " *** Done ***" << endl << endl; } return exitCode; }
void CCameraBasler::continuousAcquireHardwareTrigger(int index) { IplImage *img; CGrabResultPtr ptrGrabResult; std::pair<char*,IplImage*> pairTemp; unsigned int indiceImagenes = 0; std::ofstream myfile; std::string gpsFileName= folderName; gpsFileName.append("\\gps.txt"); myfile.open (gpsFileName.c_str()); while (m_cameraState == CVCCameraInterface::GRABBING) { try { #ifdef DEBUG_PERFORMANCE Timer timer; Timer timer2; timer2.start(); #endif m_Camera->RetrieveResult(INFINITE, ptrGrabResult, TimeoutHandling_Return); if (ptrGrabResult->GrabSucceeded()) { // Convert pylon image to opencv image format img = cvCreateImage(cvSize(ptrGrabResult->GetWidth(),ptrGrabResult->GetHeight()),8,3); #ifdef DEBUG_PERFORMANCE timer.start(); #endif CBaslerFormatConverter::convertCOLORtoCVRGB(img,ptrGrabResult); #ifdef DEBUG_PERFORMANCE timer.stop(); std::cout << "[DEBUG BASLER] "<< timer.getElapsedTimeInMilliSec() << " ms In convert from Basler to Opencv\n"; #endif #ifdef DEBUG_SHOWSNAPS cvShowImage("Visible",img); cvWaitKey(1); #endif // Construct image file name char *str = new char[200]; std::string isoString = boost::posix_time::to_iso_string(boost::posix_time::microsec_clock::universal_time()); std::string date = isoString.substr(0,8); std::string time = isoString.substr(9,20); myfile << date << "T,"<< time<<"Z,"<<CVCCameraInterface::GPSInfo::Instance()->gpsInfo; sprintf(str,"%s\\Visible_%d.bmp",folderName.c_str(),indiceImagenes); // Put name and image in a pair pairTemp.first = str; pairTemp.second = img; // Enqueue pair of <nameOfFileToSave,image> in the SaveImageProcess // Is a singleton thread save design to enqueue and process all images. // Index refers to the "id" of the queue. Each queue will have a thread. CVCCameraInterface::SaveImagesProcess::Instance()->Add(index,pairTemp); // Increment the number of the actual image. Used to make the filename indiceImagenes++; #ifdef DEBUG_PERFORMANCE std::cout << "[DEBUG BASLER] "<< indiceImagenes << " imagenes tomadas"<<std::endl; std::cout << "[DEBUG BASLER] "<< timer2.getElapsedTimeInMilliSec() << " Duro el ciclo if"<<std::endl; #endif } else { std::cout << "Error: [Basler] :" << ptrGrabResult->GetErrorCode() << " " << ptrGrabResult->GetErrorDescription() << std::endl; } } catch(GenICam::GenericException &e) { std::cerr << "ERROR: [Basler] Acquiring continuous images." << e.GetDescription() << std::endl; } } myfile.close(); try { m_Camera->StopGrabbing(); } catch(GenICam::GenericException &e) { std::cerr << "ERROR: [Basler] Acquiring continuous images." << e.GetDescription() << std::endl; } }
IplImage * CCameraBasler::AcquireSingleImage() { CGrabResultPtr ptrGrabResult; IplImage *img = NULL; try { if (m_Camera == NULL) { std::cerr << "ERROR: [Basler] Camera is null." << std::endl; return NULL; } // Camera have to be configured in free running if ( m_cameraState == CVCCameraInterface::CONFIGURED ) { // Camera have to be configured in trigger software mode. If is in custom mode it's ok, otherwise will force // default trigger software configuration. if ( m_cameraStateMode != CVCCameraInterface::CUSTOM ) { if ( m_cameraStateMode != CVCCameraInterface::TRIGGER_SOFTWARE ) { configureCameraMode(CVCCameraInterface::TRIGGER_SOFTWARE); } } } // Configure strategy one by one m_Camera->StartGrabbing( GrabStrategy_OneByOne); // Wait camera to be ready for a software trigger if (m_Camera->WaitForFrameTriggerReady( 200,TimeoutHandling_ThrowException)) { // Execute software trigger m_Camera->ExecuteSoftwareTrigger(); } // Wait some time to acquire the image ::Sleep(200); // Grab the image from the pylon buffer m_Camera->RetrieveResult(5000,ptrGrabResult,TimeoutHandling_Return); // If is not grabbed succesfull if (!ptrGrabResult->GrabSucceeded()) { return NULL; } // Stop camera grabbing m_Camera->StopGrabbing(); } catch(GenICam::GenericException &e) { std::cerr << "Error: [Basler] An exception occurred acquiring single image." << std::endl << e.GetDescription() << std::endl; return NULL; } try { // Convert pylon image to opencv image format img = cvCreateImage(cvSize(ptrGrabResult->GetWidth(),ptrGrabResult->GetHeight()),8,3); CBaslerFormatConverter::convertCOLORtoCVRGB(img,ptrGrabResult); } catch(...) { return NULL; } // Return the image in opencv format return img; }
int main(int argc, char* argv[]) { // The exit code of the sample application. int exitCode = 0; if(const char* env_p = std::getenv("PATH")) std::cout << "Your PATH is: " << env_p << '\n'; // Automagically call PylonInitialize and PylonTerminate to ensure the pylon runtime system // is initialized during the lifetime of this object. Pylon::PylonAutoInitTerm autoInitTerm; try { // Create an instant camera object with the camera device found first. CInstantCamera camera( CTlFactory::GetInstance().CreateFirstDevice()); // Print the model name of the camera. cout << "Using device " << camera.GetDeviceInfo().GetModelName() << endl; INodeMap& nodemap = camera.GetNodeMap(); // Open the camera for accessing the parameters. camera.Open(); // Get camera device information. cout << "Camera Device Information" << endl << "=========================" << endl; cout << "Vendor : " << CStringPtr( nodemap.GetNode( "DeviceVendorName") )->GetValue() << endl; cout << "Model : " << CStringPtr( nodemap.GetNode( "DeviceModelName") )->GetValue() << endl; cout << "Firmware version : " << CStringPtr( nodemap.GetNode( "DeviceFirmwareVersion") )->GetValue() << endl << endl; // Camera settings. cout << "Camera Device Settings" << endl << "======================" << endl; // Set the AOI: // Get the integer nodes describing the AOI. CIntegerPtr offsetX( nodemap.GetNode( "OffsetX")); CIntegerPtr offsetY( nodemap.GetNode( "OffsetY")); CIntegerPtr width( nodemap.GetNode( "Width")); CIntegerPtr height( nodemap.GetNode( "Height")); // Access the PixelFormat enumeration type node. CEnumerationPtr pixelFormat( nodemap.GetNode( "PixelFormat")); // Remember the current pixel format. String_t oldPixelFormat = pixelFormat->ToString(); cout << "Old PixelFormat : " << oldPixelFormat << endl; // Set the pixel format to Mono8 if available. if ( IsAvailable( pixelFormat->GetEntryByName( "Mono8"))) { pixelFormat->FromString( "Mono8"); cout << "New PixelFormat : " << pixelFormat->ToString() << endl; } // The parameter MaxNumBuffer can be used to control the count of buffers // allocated for grabbing. The default value of this parameter is 10. camera.MaxNumBuffer = 20; // Start the grabbing of c_countOfImagesToGrab images. // The camera device is parameterized with a default configuration which // sets up free-running continuous acquisition. camera.StartGrabbing( c_countOfImagesToGrab); // This smart pointer will receive the grab result data. CGrabResultPtr ptrGrabResult; // Camera.StopGrabbing() is called automatically by the RetrieveResult() method // when c_countOfImagesToGrab images have been retrieved. while ( camera.IsGrabbing()) { // Wait for an image and then retrieve it. A timeout of 5000 ms is used. camera.RetrieveResult( 5000, ptrGrabResult, TimeoutHandling_ThrowException); // Image grabbed successfully? if (ptrGrabResult->GrabSucceeded()) { // Access the image data. cout << "SizeX: " << ptrGrabResult->GetWidth() << endl; cout << "SizeY: " << ptrGrabResult->GetHeight() << endl; const uint8_t *pImageBuffer = (uint8_t *) ptrGrabResult->GetBuffer(); cout << "Gray value of first pixel: " << (uint32_t) pImageBuffer[0] << endl << endl; #ifdef PYLON_WIN_BUILD // Display the grabbed image. Pylon::DisplayImage(1, ptrGrabResult); #endif } else { cout << "Error: " << ptrGrabResult->GetErrorCode() << " " << ptrGrabResult->GetErrorDescription() << endl; } } } catch (GenICam::GenericException &e) { // Error handling. cerr << "An exception occurred." << endl << e.GetDescription() << endl; exitCode = 1; } // Comment the following two lines to disable waiting on exit. cerr << endl << "Press Enter to exit." << endl; while( cin.get() != '\n'); return exitCode; }