Comments (8)
@lyyiangang , the openvino model is given for a Mean absolute error of 6.95°, and is not very stable.
There may be solutions : make annotations on 3D models for training, using the Iris MediaPipe model, ...
You can join us to discuss it on the #eye-saccades channel of the DepthAI Discord:
https://discord.gg/HBNhaFc4
from depthai-experiments.
Great question. So I don't immediately know how to do this but I bet @LCTyrell likely does. He's done a decent amount of work in this direction.
Edit, meant to link this:
https://github.com/LCTyrell/Object-looked-at_Estimation
from depthai-experiments.
Hi,
The description of the model is here:
https://docs.openvinotoolkit.org/latest/omz_models_intel_gaze_estimation_adas_0002_description_gaze_estimation_adas_0002.html
The NNet does not give you the pitch and yaw angles but :
" The network outputs 3-D vector corresponding to the direction of a person’s gaze in a Cartesian coordinate system in which z-axis is directed from person’s eyes (mid-point between left and right eyes’ centers) to the camera center, y-axis is vertical, and x-axis is orthogonal to both z,y axes so that (x,y,z) constitute a right-handed coordinate system."
I hope this will help you ;-)
from depthai-experiments.
Thanks @LCTyrell !
from depthai-experiments.
@LCTyrell Thanks very much for your reply. I want to use this model to estimate what the subject is looking. E.g. for a driver in a car, I want to know what the driver is looking at, front windshield, rear-view mirror or or right window rear-view? I think if the gaze vector and these gaze target are mapped to the same coordinate system, I can check the intersection of gaze ray and these objects to figure this problem. Then the question is how we can convert the gaze vector to the real world / camera coordinate system ? Is there any paper for your gaze model? Thanks very much.
from depthai-experiments.
OpenVino Gaze Estimate is a custom model of Intel. There is no much more information than the link I have provided above.
To make the conversion, you will have to recover your old (or not) courses of mathematics ;-). I'm pretty sure there is a lot of ressources for that on the web.
(ex: https://math.stackexchange.com/questions/1352632/change-from-one-cartesian-co-ordinate-system-to-another-by-translation-and-rotat )
To check if the result is good, 3D visualization (such as the 3D pose estimation) may be necessary.
If you are not in a hurry, I will work on that and put it on Github ... (but it will take time).
from depthai-experiments.
Thanks much for your reply. In fact, I am studying gaze vector estimation for Driver Monitor System. And have done many works on it by now, I have tried many gaze methods/dataset, e.g. MPII gaze, UnityEyes gaze, and also try the opengaze code, but there are not very stable in real scenario, zig-zagging often appears, accuracy can not reach 4~5 deg in fact. Waiting for your more open source code.
from depthai-experiments.
Good to know on the accuracy. IIRC, it was trained on only 50 people. So training on more will likely improve the accuracy. The hard part on this is I think getting accurate truth data.
from depthai-experiments.
Related Issues (20)
- Running YOLOv8 model on OAK D Pro Wide camera. I keep getting a not enough memory error. HOT 2
- OAK-D-SR Camera save color image and depth image by record.py HOT 3
- [gen2-yolo/device-decoding] Visualization doesn't show camera footage or detections HOT 2
- To add mesh generation for gen2-pointcloud and Multiple camera HOT 2
- How to visualize when I record data with gen2-record-replay?
- Why pylance is complaining about NNData(), node, and ImageManipConfig() call in script.py from gen2-gaze-estimation project?(Solved) HOT 2
- Head posture estimation module HOT 2
- segmentation with yolov5
- Head pose estimation HOT 2
- How to load external video or webcam feed and process it on device side HOT 1
- New recommendations on head pose estimation model. HOT 1
- Support for YOLOP HOT 2
- Depth information in deep learning model HOT 3
- Camera extrinsic parameter HOT 1
- Issue with installing requirements for RTSP streaming HOT 1
- Deploy yolov8 on OAK-D HOT 6
- Challenges Installing PyGObject for OKA Camera RTSP Streaming HOT 1
- Can we specify the layers you are interested in for a custom nn ? HOT 8
- Yolo v8 on OAK-D camera HOT 1
- gen2-pedestrian-reidentification freezes while person is detected HOT 3
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from depthai-experiments.