Kinect SDK for finger detection? - c#

I'm a student from Nanyang Technological University (NTU), Singapore. And currently developing a project using Kinect SDK.
And my question is, anyone know how to develop a program to detect a finger (or fingertip) using Kinect SDK, or maybe even some possible reference codes. Anyway, I also tried to search on google, but the only reference I got is using Open NI, instead of Kinect SDK.
Thanks and Regards

I was looking into that myself, although haven't gone deep into it.
OpenNI has some constants for finger tip/wrist detection, but that's not implemented yet,
but that's not an option for your setup anyway.
Here's a list of resources that hopefully will get you started:
MIT CSAIL Hand detection
FORTH ICS - Efficient model-based 3D tracking of hand articulations using Kinect
Candescent NUI project
Other random videos

Related

Kinect touchable Surface

I have a programming project using Kinect Xbox one sensor. The project is mainly about turning any surface into an interactive touchable screen. I have collected all the hardware including the projector. In addition, I have done my research and downloaded the related packets such as Visual Studio in order to start coding in C#.
So, my question here:
Is there any any library that I could use which may facilitate me to determine the angles/depth of the surface?
Plus, I don't have a fully vision of the steps which need to be done for the next steps, so I would really appreciate it if there is anyone could draw me a small map for me for this project.
If you have trouble with getting started with kinect go through this
Quick start series
and you also might want to capture the depth of objects. For that try to use Kinect's depth image streams and the sdk itself does not provide much happy methods. You will have to do some image processing on that gray scaled depth stream. Then you can find the edges of a single object in different depths.

Kinect 2 skinning a model

I am looking at using the Kinect version 2 for PC to control an avatar, i.e. apply a skin to the body basics example provided by Microsoft.
The avatar isn't to be used in a game or anything at this point. I simply plan to be able to control the avatar using the Kinect 2 and record to .avi. I have already implemented a demo where I can record the Kinect footage of the skeleton directly to .avi.
After searching I found that it was possible to control an avatar with the original Kinect using XNA game studio. As XNA game studio seems a bit out dated now and I'm unsure if it would be compatible with version 2, would there be an alternative approach? Does anybody know of a project where someone has done this already? Or can anyone suggest a good approach to take to do this?
I have also seen some articles referring to a plugin that lets you use Unity with Kinect 2 but from what I can see I don't believe this is publicly available?
Many thanks for any help.
Yes, the unity plug in works great and its the best out there. the creator its call rumen, he will help you if you need it help but ofcourse you have to pay the plugin 1st. its worth it, it makes kinect way simpler.

Kinect SDK 1.7 - How to work with both hands

I´m trying to work with Kinect SDK 1.7 in WPF and C#. I found some tutorials, but nowhere I cannot find how to work with both hands at once (in my Kinect Region i only see just one hand at any time).
Please can you help me, how to do that? I´d like to this this for zooming in maps!
Thank you very much!
I don't think that making Zoom is easy but there is a way of showing the two hands. You have to go in the source code of the dll and make a small change.
http://social.msdn.microsoft.com/Forums/en-US/7e00d05d-8b97-4838-b3d6-76a71c4d3166/how-to-use-two-hand-cursors-simultaneously-using-kinect-interactions

Interacting with avatar using Kinect and Unity

I want to move the avatar based on the movement the player using kinect and Unity, are there any good tutorials?
We are using unity and Kinect interface to create a simple application. Based on the movement of the player we need to move the avatar.
We are supposed to use Unity with GAKUNITY, No OpenNI or any third party tools.
Are there any good tutorials for GakUnity with Kinect?
GAK means Gadget Accelerator Kit
We just want to move any avatar with player movement in front of kinect interface. Even help in hand movement also highly appreciated.
You can also share useful links or books regarding to unity and kinect programming.
Custom Kinect Gesture Recognition using OpenNI and Unity3Dkinect.dashhacks.com
A tipster has tipped a lil tipple my way in the form of custom gesture recognition for the Kinect using OpenNI and Unity 3D. What this allows you to do is create you own custom gesture that are recognized by the kinect through the software interface.
Unity and Kinect tutorialnightmarekitty.com
For this chapter, we are going to be using a very popular game engine called Unity. By integrating OpenNI, Nite and Sensor Kinect into Unity we will control a 3d character and multiple user interfaces. After we cover the main Components, we will build an example of each from the bottom up.
Kinect Wrapper Example Projectwiki.etc.cmu.edu
This scene shows you how a skeleton is generated / tracked by placing spheres at each of the bones tracked by the Kinect, and how to use kinect to control your model. Use this to get a feel for what the Kinect is capable of. It also shows you how to prepare your GameObjects
I am not familiar with Unity 3d,but you can try using kinect with XNA. At the latest version of Kinect for windows SDK 1.5 and Developer Tookit 1.5.1,there is a sample demonstrating how to interact with 3d avatar using Kinect and XNA,you can find more information on http://msdn.microsoft.com/en-us/library/jj131041
I have no idea about GAKUNity, but you can use zigfu plugin for unity3d and do your project. It has got sample codes and character which you can use for reference, and the sample codes are commented nicely .
You can just import your bone, drag your bones on the zigskeleton script . just read through th zig skeleton script and you will understand the function and structures they are using.
And you can even refer the MEET THE KINECT book, they have explained about using zigfu in unity .

human body border detection using .NET

there is any body know any library that enable us to detect the body border or help me notify to body movment according to stream of images
Microsoft has such a library, called Kinect.
However, I rather doubt they will license it to you.
There is not always a library for everything, especially if the problem is difficult to precisely define and on the cutting edge of technology.
The basic element of this sort of processing is Blob Detection.
There are examples in the wild for blob detection algorithms. There are libraries too, but I'm not sure about free ones.
See this:
http://www.aforgenet.com/articles/hands_gesture_recognition/
Did you look at open CV ? It's an open-source framework for image processing so it has a lot of helpful algorithms for contour detection, filtering, hull etc...
Also OpenNI was just released. It's a framework for natural interaction. For now it's mostly set up to work with depth sensors such that Kinect and PrimeSense sensors, but in the not-so-far future it might be able to deal with other type of hardware, maybe standard video cameras...
Link to MSFT SDK site
Update:
Just in from ZD.NET "Microsoft readies hobbyist Kinect development kit; promises commercial version later" link
Prior Post
The makers/OEM of the Kinect is PrimeSense. They make both the kinect and an independent devkit. They recently released open source drivers that are compatible with both XBox's Kinect and the smaller DevKit. You might as well get the drivers straight from the source. I'm not sure what value a 3rd party library (openKinect) will offer.
Get the unified Kinect and DevKit drivers here: http://www.openni.org/ Windows and Unix support is available.

Categories