When installing the SDK, remember the path you install to. For example, 'C: Program Files Azure Kinect Body Tracking SDK 1.0.0'. You will find the samples referenced in articles in this path. Body tracking samples are located in the body-tracking-samples folder in the Azure-Kinect-Samples repository. You will find the samples referenced.

  1. Kinect Sdk 3
  2. Kinect Sdk 2
  3. Kinect Sdk Linux
  4. Kinect Sdk 1.8 Vs 2.0
  • Install Kinect for Windows Runtime 2.0. Install Kinect for Windows SDK 2.0. Restart your PC. Once you have completed these steps, you can open the camera app on Windows 10 to verify that.
  • Azure Kinect is a cutting-edge spatial computing developer kit with sophisticated computer vision and speech models, advanced AI sensors, and a range of powerful SDKs that can be connected to Azure cognitive services.
  • Kinect for Windows SDK 2.0. Build desktop applications for Windows 10 or sell your Kinect v2 UWP apps in the Microsoft Store to take advantage of unique Windows Runtime features and expand distribution. GET THE KINECT FOR WINDOWS SDK.
  • Use the Kinect for Windows SDK to build applications with C, C# or Visual Studio Basic by using Microsoft Visual Studio 2010. What availability of this product will do is facilitate the ability of developers and commercial customers to explore the limitless opportunities to transform products, processes, and businesses.
  • Select a Lab:

Welcome to the Kinect 2 Hands on Labs!
This series will show you how to build a Windows 8.1 store app which uses almost every feature of the Kinect 2. The lessons in this series work the best when completed in order.
You can download a master copy of the complete app and all labs and referenced libraries through the github links on the left.
Or if you know a bit about development with the Kinect 2 already, you can skip to a particular lab by navigating to it at the top of the page. The running codebase is available through a link at the bottom of each page, which is complete and runnable as if you have just finished that lab.
If you have any suggestions or would like to report any bugs, please leave some feedback on the Kinect Tutorial GitHub Issues page.
Enjoy the labs and have fun!

SystemRequirements

The target application is a Windows 8.1 Store application forx64 machines (not ARM or x86) so a machine with an x64 architecture isrequired for development.Supported Operating Systems and Architectures

  • Windows 8.1
  • Microsoft Visual Studio® 2013 Community Edition with Update 2 orhigher.
  • Microsoft .NET Framework 4.5.
  • The Kinect 2 SDK (Currently at build 1409)
  • Kinect 2.0 for Windows device.
  • USB 3.0

Debugging

Debugging the Kinect 2 requires that you meet the system requirements.
If you are unsure that the Kinect is plugged in properly, you can check a light indicator on the power box of the unit (the box which comes from the single cable in the Kinect 2) and results in power and USB 3.0 connections.

  • If the light on the power-box is Orange then something is wrong with either the power, Kinect 2, or USB3.0 connection.
  • If the light is White then the Kinect is correctly registered with windows as a device to be used.
The Kinect 2.0 itself does not turn on until it is required by an application. Kinect sdk 1.8 vs 2.0

Azure Kinect Examples for Unity, v1.16 is a set of Azure Kinect (aka ‘Kinect for Azure’, K4A) examples that use several major scripts, grouped in one folder. The package currently contains over thirty five demo scenes. Apart of the Azure Kinect sensor (aka K4A), the K4A-package supports the “classic” Kinect-v2 (aka Kinect for Xbox One) sensor, as well as Intel RealSense D400-series sensors.

The avatar-demo scenes show how to utilize Kinect-controlled avatars in your scenes, gesture demo – how to use discrete and continuous gestures in your projects, fitting room demos – how to overlay or blend the user’s body with virtual models, background removal demo – how to display user silhouettes on virtual background, point cloud demos – how to show the environment or users as meshes in your scene, etc. Short descriptions of all demo-scenes are available in the online documentation.

This package works with Azure Kinect (aka Kinect for Azure, K4A), Kinect-v2 (aka Kinect for Xbox One) and Intel RealSense D400-series sensors. It can be used with all versions of Unity – Free, Plus & Pro.

How to run the demo scenes:
1. (Azure Kinect) Download and install the latest release of Azure-Kinect Sensor SDK. The download link is below. Then open ‘Azure Kinect Viewer’ to check, if the sensor works as expected.
2. (Azure Kinect) Follow the instructions on how to download and install the latest release of Azure-Kinect Body Tracking SDK and its related components. The link is below. Then open ‘Azure Kinect Body Tracking Viewer’ to check, if the body tracker works as expected.
3. (Kinect-v2) Download and install Kinect for Windows SDK 2.0. The download link is below.
4. (RealSense) Download and install RealSense SDK 2.0. The download link is below.
5. Import this package into a new Unity project.
6. Open ‘File / Build settings’ and switch to ‘PC, Mac & Linux Standalone’, Target platform: ‘Windows’ & Architecture: ‘x86_64’.
7. Make sure that ‘Direct3D11’ is the first option in the ‘Auto Graphics API for Windows’-list setting, in ‘Player Settings / Other Settings / Rendering’.
8. Open and run a demo scene of your choice from a subfolder of the ‘AzureKinectExamples/KinectDemos’-folder. Short descriptions of all demo-scenes are available in the online documentation.

* The latest Azure Kinect Sensor SDK (v1.4.1) can be found here.
* The latest Azure Kinect Body Tracking SDK (v1.1.0) can be found here.
* Older releases of Azure Kinect Body Tracking SDK can be found here.
* Instructions how to install the body tracking SDK can be found here.

2.0

* Kinect for Windows SDK 2.0 can be found here.
* RealSense SDK 2.0 can be found here.

Downloads:
* The K4A-asset may be purchased and downloaded in the Unity Asset store. All updates are and will be available to all customers, free of any charge.
* If you’d like to try the free version of the K4A-asset, you can find it here.
* If you’d like to utilize the Intel RealSense sensor with Cubemos body tracking, please look at this tip.
* If you’d like to utilize the LiDAR sensor on your iPhone-Pro or iPad-Pro as a depth sensor, please look at this tip.

Kinect Sdk 3

Free for education:
The package is free for academic use. If you are a student, lecturer or university researcher, please e-mail me to get a free copy of the K4A-asset directly from me.

One request:
Please don’t share this package or its demo scenes in source form with others, or as part of public repositories, without my explicit consent.

Unity

Kinect Sdk 2

Kinect sdk 2.2

Kinect Sdk Linux

Documentation:
* The basic documentation is in the Readme-pdf file, in the package.
* The K4A-asset online documentation is available here and as pdf.
* Many K4A-package tips, tricks and examples are available here.

Troubleshooting:
* If you can’t upgrade the K4A-package in your project to the latest release, please go to ‘C:/Users/{user-name}/AppData/Roaming/Unity/Asset Store-5.x’ on Windows or ‘/Users/{user-name}/Library/Unity/Asset Store-5.x’ on Mac, find and delete the currently downloaded package, and then try again to download and import it.
* If Unity editor freezes or crashes at the scene start, please make sure the path where the Unity project resides does not contain any non-English characters.
* If you get syntax errors in the console like “The type or namespace name ‘UI’ does not exist…”, please open the Package manager (menu Window / Package Manager) and install the ‘Unity UI’ package. The UI elements are extensively used in the K4A-asset demo scenes.
* If you get “‘KinectInterop.DepthSensorPlatform’ does not contain a definition for ‘DummyK2′” in the console, please delete ‘DummyK2Interface.cs’ from the KinectScripts/Interfaces-folder. This dummy interface is replaced now with DummyK4AInterface.cs.
* If the Azure Kinect sensor cannot be started, because StartCameras()-method fails, please check again #6 in ‘How to run the demo scenes‘-section above.
* If you get a ‘Can’t create the body tracker’-error message, please check again #2 in ‘How to run the demo scenes‘-section above. Check also, if the Body Tracking SDK is installed in ‘C:Program FilesAzure Kinect Body Tracking SDK’-folder.
* If the body tracking stops working at run-time or the Unity editor crashes without notice, update to the latest version of the Body tracking SDK. This is a known bug in BT SDK v0.9.0.
* Regarding RealSense: If you’d like to try its integration with Cubemos body tracking SDK, please look at this tip.
* If there are errors like ‘Shader error in [System 1]…’, while importing the K4A-asset, please note this is not really an error, but shader issues due to missing HDRP & VFX packages. You only need these packages for the Point-cloud demo. All other scenes should be started without any issues.
* If there are compilation errors in the console, or the demo scenes remain in ‘Waiting for users’-state, make sure you have installed the respective sensor SDKs and the other needed components. Please also check, if the sensor is connected.

What’s New in Version 1.16:
1. Added support for Azure Kinect Body Tracking SDK v1.1.0 (big thanks to 葛西浩!).
2. Added optional bone colliders to the AvatarController-component.
3. Updated BackgroundUserBodyImage- & BackgroundColorCamUserImage-components to support individual user indexes (thanks to Tepat Huleux).
4. Fixed SceneMeshDemo and BackgroundRemovalDemo2-scenes to limit the space in camera coordinates (thanks to Tomas Durkin).
5. Fixed the dst/cst transfer issue in case of KinectNetServer running with ARKit sensor interface.

Kinect Sdk 1.8 Vs 2.0

Videos worth more than 1000 words:
Here is a holographic setup, created by i-mmersive GmbH, with Unity 2019.1f2, Azure Kinect sensor and “Azure Kinect Examples for Unity”: