Maybe I'm dumb but i tried reading the answers you guys had here, but I couldn't figure out how to do it. So I got a game (a Stargate shooter) where you drive a space ship and you gotta destroy all of your opponents. The problem I'm having is I don't really know how to create a map and how to move my camera around it according to the ship's position. Can anyone help?
You have probably resolved this yourself already, but if not then I highly recommend going over to David Amador's page on the subject. As an added bonus, David is active on the site and replies directly to your questions (within 10 minutes usually).
Related
Anyone know a binary that exports the output of the coordinates? I am looking for something related to several days, I have already visited dozens of sites, downloaded dozens of projects and researched dozens of tutorials, but I can't find anything that really works, I'm trying to create an application similar to VTuber (Virtual Youtubers) in Construct 2 where Head Tracking is crucial to make movement work, I understand the complexity of developing such an engine, but I also imagine that there is something easily accessible out there on the internet, for example...
I found this website that in the lower right corner has the information I would need, but it is a website, it is not quite a binary that I keep monitoring the output to make the character animate inside Construct 2...
Site: https://www.visagetechnologies.com/HTML5/latest/Samples/ShowcaseDemo/ShowcaseDemo.html
I also found a very cool project, but it is also a website, although I really liked it because it is very light running.
Site: https://www.auduno.com/clmtrackr/examples/clm_emotiondetection.html
I also found a project called "ARKitFaceTracker" but it was made for Unity and the level of complexity of the project is discouraging for me.
Currently a project already exists that uses "ARKitFaceTracker" is PrprLive, it was made at Unity and it would be perfect if you exported the horns in a text file, so that you could try to make a lighter version and in my own way, but I just I know how to program in VB.Net and a little bit of C#, but this "Face Camera" was basically what I wanted, but in a way I just used it as "engine", to get the direction coordinates of the head.
In short: basically what I'm looking for is a binary executable that can export the only the movement (left, right, up and down) coordinates of this "Head Tracking" so I can pull this "Output" (txt or json) through Construct 2 and make it animate in real time. I really wanted to make it happen, mainly because I have been researching and searching the internet for several days and it has been a few hours since I crashed and I came here wanting to know if you could help me to continue, because I really don't know what else do, thank you very much.
It can be easier if you modify from projects like this:
https://github.com/1996scarlet/OpenVtuber
https://github.com/virtuber/openvtuber
Both these solutions separate the face tracking module and the visualization module. It should be easy for you to set up the python servers and request coordinates from them.
I want different images to be displayed from different point of view. For the whole concept explaination please look at the images. they explain my idea/query!
As in the first image you see that there are three people at different angle looking at the monitor. Now i want the webcam to track the eyes and show the particular defined image to the user> For example: If user is at 45 degree angle then show image1.png
Depending upon the user's prespective of watching. The computer should show the image.
(the lady is the game character for representation purpose)
Can you please guide me on what steps can be taken to accomplish this? Is there any plugin available for unity that tracks faces? Please guide me
Also thanks for the compliments on my sketching skills xD
Stackoverflow is not really meant to recommend plugins, since the choice is usually opinion based so there is no exact answer.
That being said, on of the most common used API for computer vision (meaning interpreting images, including face recognition) is OpenCV, so that could be a good start for you to look at that.
And fortunately for you, there is a Unity plugin for OpenCV
It is too broad to give you more details about how it works here. You should try to make it work, and if you have a problem with your code, open a new question with the code portion that you struggle with.
PS: nice sketching skills
Perhaps easier option would be to use Kinect
(trying to detect face or eyes from that far might be shaky?)
With Kinect you can get skeletons for multiple people, and getting the angle between target and those kinect avatars would be easy.
If there is no space to put kinect in good position,
could consider placing it on the ceiling above (and then use depth data only to detect people in its view)
Only issue is that apparently Microsoft has stopped Windows kinect support,
so you would need to find 2nd hand versions.. (Unity Asset store still has some kinect plugins and examples available)
https://www.polygon.com/2018/1/2/16842072/xbox-one-kinect-adapter-out-of-stock-production-ended
Or look for kinect alternatives that work with unity, try RealSense cameras:
https://www.intel.sg/content/www/xa/en/architecture-and-technology/realsense-overview.html
I am looking at using the Kinect version 2 for PC to control an avatar, i.e. apply a skin to the body basics example provided by Microsoft.
The avatar isn't to be used in a game or anything at this point. I simply plan to be able to control the avatar using the Kinect 2 and record to .avi. I have already implemented a demo where I can record the Kinect footage of the skeleton directly to .avi.
After searching I found that it was possible to control an avatar with the original Kinect using XNA game studio. As XNA game studio seems a bit out dated now and I'm unsure if it would be compatible with version 2, would there be an alternative approach? Does anybody know of a project where someone has done this already? Or can anyone suggest a good approach to take to do this?
I have also seen some articles referring to a plugin that lets you use Unity with Kinect 2 but from what I can see I don't believe this is publicly available?
Many thanks for any help.
Yes, the unity plug in works great and its the best out there. the creator its call rumen, he will help you if you need it help but ofcourse you have to pay the plugin 1st. its worth it, it makes kinect way simpler.
i am a 3rd year computer science student, and my seminar(graduation project) is in the next semester, so i started to think about it and i thought that a face recognition system would be a great idea, so i have two questions:
1- how can i read a picture programmatically, i mean if i have a picture, how can read the pixels and colors in order to recognize the face later on?
2- do you know any useful ebooks or resources that might help me in this topic cuzz i am really interested about it, and i hope that i can implement it.
thank you very much dudes
this is my first post and i hope you will really help me :)
Take a look at OpenCV which makes access to webcams easy! It can also help you with further picture computations.
Take a look at AForge.NET library. This library has very usefull API for computer vision. Take a look also on articles writen by author of this library - Andrew Kirillov. He wrote about motion detection algorithms and hand gesture recognition.
i have been searching for long in order to find how i can place a sphere/cube around my playfield so ill have an effect similar to sky. I have seen SkySpheres/Boxes but couldnt implement them . Im looking for a rather simple solution since my play field is really small and i just need something to replace the Clear(Color.Black) .I dont care about collision i just need it to be there with a texture on . Thanks !
You should take a look at Riemers XNA Tutorial
HERE is the post on the skybox.