Behavioral analysis toolkit
for edge devices

Advanced gaze estimation, head pose tracking, and cognitive assessment toolkit for real-world applications

Production-Ready Models

State-of-the-art AI models optimized for mobile and edge devices

✓ Production Ready

6D Head Pose Estimation

State-of-the-art head pose tracking using continuous 6D rotation representation. Outperforms RepVGG-B1 baseline while using 35% fewer parameters through RepNeXt-M4 architecture.

3.91° Mean Angular Error
13.8M Parameters
35% Size Reduction
Mobile Optimized
  • Architecture: RepNeXt-M4 backbone with 6D rotation output
  • Training: AFLW2000 test set, 300W-LP training set
  • Method: Gram-Schmidt orthonormalization for SO(3)
  • Loss: Geodesic distance on rotation manifold
  • Comparison: 3.91° vs 3.97° (RepVGG-B1 20M params)

Key Innovation: Replaces Euler angles with continuous 6D representation to avoid gimbal lock, ensuring stable rotations across full ±80° yaw/pitch range.

RepNeXt-M4 6D Rotation Geodesic Loss Mobile Edge
⚙ In Development

RayNet: Multi-View Gaze System

Stereo gaze estimation architecture leveraging 9-camera multi-view consistency. Combines PANet feature pyramid with EyeFLAME module for joint 2D-3D supervision resolving depth ambiguity.

95.48px 2D Iris Error
14.89cm 3D Eyeball Error
9 Views Multi-Camera
30.22° Gaze Angle Error
  • Dataset: GazeGene (1,008,000 images, 56 subjects)
  • Architecture: RepNeXt + PANet + EyeFLAME heads
  • Multi-Task: Gaze vector, depth, head pose, pupil center
  • Training: 3-phase progressive strategy (2D→3D)
  • Camera Setup: 9 synchronized views at 5m distance

Progressive Training: Phase 1 focuses on 2D projection (weight=1.0), Phase 2 introduces 3D (weight=0.1), Phase 3 balances both (weights=0.5) for optimal convergence.

Stereo Vision PANet FPN Multi-Task MTL GazeGene
⚙ Work in Progress

EyeFLAME: 3D Eye Structure

FLAME-inspired geometric module integrated with RayNet for anatomical eye reconstruction using multiple camera streams. Predicts 3D eyeball structures with true perspective projection leveraging multi-view consistency.

200 Iris Landmarks
15.11cm 3D Iris Error
9 Cameras Multi-View
3D Perspective
  • Architecture: Module within RayNet for multi-camera processing
  • Output: Eyeball centers, pupil centers, 100 iris points per eye
  • Method: 3D perspective model with multi-view geometry
  • Supervision: Joint 2D-3D loss with angular constraints
  • Geometry: Subject-specific kappa angles and anatomical radii

Multi-View Approach: Utilizes RayNet architecture to process 9 synchronized camera streams, resolving depth ambiguity through geometric triangulation rather than single-view weak perspective.

RayNet Integration 3D Perspective Multi-Camera FLAME-Style
⚙ In Development

ARGaze Gaze Point Estimation

Single-camera gaze point estimation on target plane using RepNeXt-M3. Training on ARGaze dataset with fixed 300cm depth assumption for desktop interaction scenarios.

9.66° Best MAE
7.8M Parameters
2000 Samples/Subject
30 Epochs
  • Dataset: ARGaze (1.3M image pairs from 25 subjects)
  • Architecture: RepNeXt-M3 backbone (lightweight)
  • Task: 2D gaze point projection on fixed-depth plane
  • Resolution: 32×32 eye images at 50fps
  • Application: Desktop/laptop gaze tracking

Ray Casting: Gaze point calculated via ray intersection: pupil center + t·gaze_direction where t = (d_plane - p_z) / v_z for fixed depth d=300cm.

RepNeXt-M3 ARGaze Single View Desktop Tracking

Built for Real-World Applications

From mobile devices to cognitive analysis systems

🎯

Mobile Optimized

Lightweight architectures designed for edge deployment with minimal latency and maximum accuracy.

🧠

Cognitive Assessment

Analyze iris patterns, pupil dynamics, and gaze behavior for cognitive load estimation.

📐

Geometric Precision

True 3D reconstruction of eye anatomy with accurate depth perception and multi-view consistency.

Real-Time Processing

Optimized inference pipelines for real-time applications in VR, AR, and interactive systems.

🔬

Research Validated

Models trained on state-of-the-art datasets including GazeGene, ETH-XGaze, and ARGaze.

🎨

Modular Design

Flexible architecture supporting multiple backbones and easy integration into existing pipelines.

Research Foundation

Built on cutting-edge computer vision and deep learning research

Key Innovations

EyeFlame combines multiple breakthrough techniques to achieve superior performance:

PyTorch Computer Vision RepNeXt Multi-Task Learning