/* Copyright (c) 2013-2015 Stanislaw Halik * * Permission to use, copy, modify, and/or distribute this software for any * purpose with or without fee is hereby granted, provided that the above * copyright notice and this permission notice appear in all copies. */ #include "ftnoir_tracker_aruco.h" #include "opentrack/plugin-api.hpp" #include #include #include #include #include "opentrack-compat/camera-names.hpp" #include "opentrack-compat/sleep.hpp" #include "opentrack-compat/pi-constant.hpp" #include #include #include #include #include #include #include struct resolution_tuple { int width; int height; }; static resolution_tuple resolution_choices[] = { { 640, 480 }, { 320, 240 }, { 320, 200 }, { 0, 0 } }; Tracker::Tracker() : stop(false), layout(nullptr), videoWidget(nullptr), obj_points(4), intrinsics(decltype(intrinsics)::eye()), dist_coeffs(decltype(dist_coeffs)::zeros()), rmat(decltype(rmat)::eye()), roi_points(4), last_roi(65535, 65535, 0, 0), freq(cv::getTickFrequency()), // XXX change to timer.hpp cur_fps(0) { // param 2 ignored for Otsu thresholding. it's required to use our fork of Aruco. detector.setThresholdParams(5, -1); detector.setDesiredSpeed(3); detector._thresMethod = aruco::MarkerDetector::FIXED_THRES; } Tracker::~Tracker() { stop = true; wait(); if (videoWidget) delete videoWidget; if(layout) delete layout; // fast start/stop causes breakage portable::sleep(1000); camera.release(); } void Tracker::start_tracker(QFrame* videoframe) { videoframe->show(); videoWidget = new cv_video_widget(videoframe); QHBoxLayout* layout_ = new QHBoxLayout(); layout_->setContentsMargins(0, 0, 0, 0); layout_->addWidget(videoWidget); if (videoframe->layout()) delete videoframe->layout(); videoframe->setLayout(layout_); videoWidget->show(); start(); for (int i = 0; i < 6; i++) pose[i] = 0; layout = layout_; } #define HT_PI OPENTRACK_PI void Tracker::getRT(cv::Matx33d& r_, cv::Vec3d& t_) { QMutexLocker l(&mtx); r_ = r; t_ = t; } bool Tracker::detect_with_roi() { if (last_roi.width > 1 && last_roi.height > 1) { float min = std::min(1.f, std::max(.01f, size_min * grayscale.cols / last_roi.width)); float max = std::max(.01f, std::min(1.f, size_max * grayscale.cols / last_roi.width)); detector.setMinMaxSize(min, max); cv::Mat grayscale_ = grayscale(last_roi); detector.detect(grayscale_, markers, cv::Mat(), cv::Mat(), -1, false); if (markers.size() == 1 && markers[0].size() == 4) { auto& m = markers[0]; for (unsigned i = 0; i < 4; i++) { auto& p = m[i]; p.x += last_roi.x; p.y += last_roi.y; } return true; } } last_roi = cv::Rect(65535, 65535, 0, 0); return false; } bool Tracker::detect_without_roi() { detector.setMinMaxSize(size_min, size_max); detector.detect(grayscale, markers, cv::Mat(), cv::Mat(), -1, false); return markers.size() == 1 && markers[0].size() == 4; } bool Tracker::open_camera() { int rint = s.resolution; if (rint < 0 || rint >= (int)(sizeof(resolution_choices) / sizeof(resolution_tuple))) rint = 0; resolution_tuple res = resolution_choices[rint]; int fps; switch (static_cast(s.force_fps)) { default: case 0: fps = 0; break; case 1: fps = 30; break; case 2: fps = 60; break; case 3: fps = 125; break; case 4: fps = 200; break; } QMutexLocker l(&camera_mtx); camera = cv::VideoCapture(camera_name_to_index(s.camera_name)); if (res.width) { camera.set(cv::CAP_PROP_FRAME_WIDTH, res.width); camera.set(cv::CAP_PROP_FRAME_HEIGHT, res.height); } if (fps) camera.set(cv::CAP_PROP_FPS, fps); if (!camera.isOpened()) { qDebug() << "aruco tracker: can't open camera"; return false; } return true; } void Tracker::set_intrinsics() { static constexpr double pi = OPENTRACK_PI; const int w = grayscale.cols, h = grayscale.rows; const double diag_fov = static_cast(s.fov) * pi / 180.; const double fov_w = 2.*atan(tan(diag_fov/2.)/sqrt(1. + h/(double)w * h/(double)w)); const double fov_h = 2.*atan(tan(diag_fov/2.)/sqrt(1. + w/(double)h * w/(double)h)); const double focal_length_w = .5 * w / tan(.5 * fov_w); const double focal_length_h = .5 * h / tan(.5 * fov_h); intrinsics(0, 0) = focal_length_w; intrinsics(0, 2) = grayscale.cols/2; intrinsics(1, 1) = focal_length_h; intrinsics(1, 2) = grayscale.rows/2; } void Tracker::update_fps(double alpha) { std::uint64_t time = std::uint64_t(cv::getTickCount()); const double dt = std::max(0., (time - last_time) / freq); last_time = time; cur_fps = cur_fps * alpha + (1-alpha) * (fabs(dt) < 1e-6 ? 0 : 1./dt); } void Tracker::draw_ar(bool ok) { if (ok) { const auto& m = markers[0]; for (unsigned i = 0; i < 4; i++) cv::line(frame, m[i], m[(i+1)%4], cv::Scalar(0, 0, 255), 2, 8); } char buf[32]; ::snprintf(buf, sizeof(buf)-1, "Hz: %d", (int)(unsigned short)cur_fps); buf[sizeof(buf)-1] = '\0'; cv::putText(frame, buf, cv::Point(10, 32), cv::FONT_HERSHEY_PLAIN, 2, cv::Scalar(0, 255, 0), 1); } void Tracker::clamp_last_roi() { if (last_roi.x < 0) last_roi.x = 0; if (last_roi.y < 0) last_roi.y = 0; if (last_roi.width < 1) last_roi.width = 1; if (last_roi.height < 1) last_roi.height = 1; if (last_roi.x >= color.cols-1) last_roi.x = color.cols-1; if (last_roi.width >= color.cols-1) last_roi.width = color.cols-1; if (last_roi.y >= color.rows-1) last_roi.y = color.rows-1; if (last_roi.height >= color.rows-1) last_roi.height = color.rows-1; last_roi.width -= last_roi.x; last_roi.height -= last_roi.y; } void Tracker::set_points() { using f = float; const f hx = f(s.headpos_x), hy = f(s.headpos_y), hz = f(s.headpos_z); static constexpr float size = 100; const int x1=1, x2=2, x3=3, x4=0; obj_points[x1] = cv::Point3f(-size + hx, -size + hy, 0 + hz); obj_points[x2] = cv::Point3f(size + hx, -size + hy, 0 + hz); obj_points[x3] = cv::Point3f(size + hx, size + hy, 0 + hz); obj_points[x4] = cv::Point3f(-size + hx, size + hy, 0 + hz); } void Tracker::draw_centroid() { repr2.clear(); static const std::vector centroid { cv::Point3f(0, 0, 0) }; cv::projectPoints(centroid, rvec, tvec, intrinsics, dist_coeffs, repr2); auto s = cv::Scalar(255, 0, 255); cv::circle(frame, repr2[0], 4, s, -1); } void Tracker::set_last_roi() { roi_projection.clear(); using f = float; cv::Point3f h(f(s.headpos_x), f(s.headpos_y), f(s.headpos_z)); for (unsigned i = 0; i < 4; i++) { cv::Point3f pt(obj_points[i] - h); roi_points[i] = pt * c_search_window; } cv::projectPoints(roi_points, rvec, tvec, intrinsics, dist_coeffs, roi_projection); set_roi_from_projection(); } void Tracker::set_rmat() { cv::Rodrigues(rvec, rmat); euler = cv::RQDecomp3x3(rmat, m_r, m_q); QMutexLocker lck(&mtx); for (int i = 0; i < 3; i++) pose[i] = tvec(i) * .1; pose[Yaw] = euler[1]; pose[Pitch] = -euler[0]; pose[Roll] = euler[2]; r = rmat; t = cv::Vec3d(tvec[0], -tvec[1], tvec[2]); } void Tracker::set_roi_from_projection() { last_roi = cv::Rect(color.cols-1, color.rows-1, 0, 0); for (unsigned i = 0; i < 4; i++) { const auto& proj = roi_projection[i]; int min_x = std::min(int(proj.x), last_roi.x), min_y = std::min(int(proj.y), last_roi.y); int max_x = std::max(int(proj.x), last_roi.width), max_y = std::max(int(proj.y), last_roi.height); last_roi.x = min_x; last_roi.y = min_y; last_roi.width = max_x; last_roi.height = max_y; } clamp_last_roi(); } void Tracker::run() { cv::setNumThreads(0); using std::fabs; using std::atan; using std::tan; using std::sqrt; if (!open_camera()) return; last_time = std::uint64_t(cv::getTickCount()); while (!stop) { { QMutexLocker l(&camera_mtx); if (!camera.read(color)) continue; } cv::cvtColor(color, grayscale, cv::COLOR_RGB2GRAY); color.copyTo(frame); set_intrinsics(); update_fps(alpha_); markers.clear(); const bool ok = detect_with_roi() || detect_without_roi(); if (ok) { set_points(); if (!cv::solvePnP(obj_points, markers[0], intrinsics, dist_coeffs, rvec, tvec, false, cv::SOLVEPNP_DLS)) goto fail; if (!cv::solvePnP(obj_points, markers[0], intrinsics, dist_coeffs, rvec, tvec, true, cv::SOLVEPNP_ITERATIVE)) goto fail; set_last_roi(); draw_centroid(); set_rmat(); } else fail: // no marker found, reset search region last_roi = cv::Rect(65535, 65535, 0, 0); draw_ar(ok); if (frame.rows > 0) videoWidget->update_image(frame); } // give opencv time to exit camera threads, etc. portable::sleep(500); } void Tracker::data(double *data) { QMutexLocker lck(&mtx); data[Yaw] = pose[Yaw]; data[Pitch] = pose[Pitch]; data[Roll] = pose[Roll]; data[TX] = pose[TX]; data[TY] = pose[TY]; data[TZ] = pose[TZ]; } TrackerControls::TrackerControls() { tracker = nullptr; calib_timer.setInterval(250); ui.setupUi(this); setAttribute(Qt::WA_NativeWindow, true); ui.cameraName->addItems(get_camera_names()); tie_setting(s.camera_name, ui.cameraName); tie_setting(s.resolution, ui.resolution); tie_setting(s.force_fps, ui.cameraFPS); tie_setting(s.fov, ui.cameraFOV); tie_setting(s.headpos_x, ui.cx); tie_setting(s.headpos_y, ui.cy); tie_setting(s.headpos_z, ui.cz); connect(ui.buttonBox, SIGNAL(accepted()), this, SLOT(doOK())); connect(ui.buttonBox, SIGNAL(rejected()), this, SLOT(doCancel())); connect(ui.btn_calibrate, SIGNAL(clicked()), this, SLOT(toggleCalibrate())); connect(this, SIGNAL(destroyed()), this, SLOT(cleanupCalib())); connect(&calib_timer, SIGNAL(timeout()), this, SLOT(update_tracker_calibration())); connect(ui.camera_settings, SIGNAL(clicked()), this, SLOT(camera_settings())); } void TrackerControls::toggleCalibrate() { if (!calib_timer.isActive()) { s.headpos_x = 0; s.headpos_y = 0; s.headpos_z = 0; calibrator.reset(); calib_timer.start(); } else { cleanupCalib(); auto pos = calibrator.get_estimate(); s.headpos_x = pos(0); s.headpos_y = pos(1); s.headpos_z = pos(2); } } void TrackerControls::cleanupCalib() { if (calib_timer.isActive()) calib_timer.stop(); } void TrackerControls::update_tracker_calibration() { if (calib_timer.isActive() && tracker) { cv::Matx33d r; cv::Vec3d t; tracker->getRT(r, t); calibrator.update(r, t); } } void TrackerControls::doOK() { s.b->save(); close(); } void TrackerControls::doCancel() { close(); } void TrackerControls::camera_settings() { open_camera_settings(tracker ? &tracker->camera : nullptr, s.camera_name, tracker ? &tracker->camera_mtx : nullptr); } OPENTRACK_DECLARE_TRACKER(Tracker, TrackerControls, TrackerDll)