Face AR enabled with our Unity face tracking
Banuba's Face AR SDK provides support for Unity 3D face tracking technology enabling you to track the face in a video stream with all the landmarks and facial expressions.
You can replicate Snapchat-like lens, AR effects, live emojis in Unity and implement them with our face mesh Unity package shipped with mask templates. The best part? It can be easily integrated with your Unity projects or used to create cross-platform augmented reality apps from scratch.
Unity Face AR Components
- Face Filters
- Virtual try on
- Avatars & Emojis
Bring face filters, 2D or 3D masks and AR effects in your Unity projects. Our plugin includes 30 face filters that you can modify per your needs. You can create the most sophisticated and realistic effects with a variety of graphical features that will keep your users entertained. Add masks with funny beards or hairstyles. Create beauty effects with vivid animations. Add virtual try on masks with glasses, hats or jewelry. Design your custom filters to add them in Unity or use the ready ones from our collection featuring over 600 AR face filters.
Virtual try on
Allow consumers to try on products in augmented reality. Our Unity Face AR plugin enables the most realistic real-time "try before you buy" experience that will help you to engage consumers and boost your revenue. The plugin includes 4 try-on packs for eyewear, headwear, hairstyles and accessories. You can use them separately to show one item at a time or combine them in mixed looks.
Avatars & Emojis
Let your users transform themselves into avatars and live emojis. With our Unity face tracking technology supporting action units, you can replace the user face with a 3D model and make it copy user facial expressions in real-time. Build avatar-based mobile apps, video chats, 3D / 2D games in Unity and deploy them across mobile. Acting like animated characters, your users will enjoy enhanced private communication and more immersive experience.
Unity Face AR Features
3D Face Mesh
We build a 3D model of the face with 3308 vertices to map face filters in real-time. The high polygon face mesh forecasts the appearance of the face in the subsequent frame increasing mask stability and precision.
You can use the face mesh debug mask featuring 86 landmarks to track face filter performance.
We support 21 action units to enable facial animation in 3D emojis and avatars that copy the user mimics and facial expressions.
Design face filters with small scenarios enabled with user facial expressions like mouth opened, smile, raised or shifted eyebrows, and opened or closed eyes.
Build and apply AR filters in Unity to multiple users simultaneously — one effect for all or different for each person. Depending on the system capabilities, 5-8 faces can be tracked on desktop and 2-3 on mobile.
Single Unity face tracking plugin that enables face filters and avatars in Windows, MacOS, iOS, Android and Web projects
You get use samples with scenes, 30 masks featuring different use cases, examples of camera usage and documentation to get started easily.
We support you all the way with our product and update our technology with new features that you get at no extra cost.