Update 3 or Nov. 16th:
Footage below of unsupervised funscript generation (green in the graph), comparing to the human made funscript (blue in the graph) for 4 randomly picked sections of the video:
Update 2 of Nov. 11th:
Still tweaking the BJ/HJ algo, but now disregards scenes that it considers irrelevant.
Model trained on 4500+ hand picked and tagged images
edit: I was curious, so I countedā¦ 30149 classes tagged in those images to train the model
Introduced Kalman filtering (againā¦ ) and fallback tracking (booooobies ).
Below a 10 random picked and unsupervised processed cuts of a video:
Another shot of the algo capabilities and inabilities at this stage, in a better resolution.
Still tweakingā¦
Though for reference, the base video is a reencoded 60 fps VR video at 1920 x 1080 and the whole inference process runs at 90fps with monitoring/displaying/logging on my Mac mini when I was barely getting 8 to 20 fps with the previous trackers solution.
Update 1 of Nov. 11th:
Made some progress, got rid of all the OpenCV tracking, now fully relying on a model I trained and algorithm to detect what is relevant to use in the frame.
This footage below is fully unsupervised and picks 4 random sections of 240 frames in a video. I still need to fix the blowjob algorithm, got a glitch in the matrix when fixing another section of the algorithm, but almost there:
Hi all,
Dreaming from being able to enjoy scripts, even basic, with whichever VR video from my collection, Iāve been working on (yet another) funscript generation helper using Python and Computer Vision techniques for the past few weeks on my free time, and I am looking for constructive feedback (to tweak, adjust, reconsider from scratch etc. the algorithm) and motivation to keep going.
I started with no coding knowledge at all, I have been testing countless Computer Vision approaches, faced disappointment and frustration, close to going batshit crazy at some point, and even lost countless hours on things as simple as the plain simple GUIā¦
And I wonāt mention the Kalman filtering (ouch, too late), nor the occlusion handling that keeps me awake at night.
So please, manage your expectations
Anyway, at this stage (I do not even know if we can call this alpha stage):
- the tool is mostly set for VR videos,
- it uses two different approaches depending on the type of scene (handjob/blowjob vs the rest),
- it could be leveraged to help scripters produce some sections of scripts before refining and fine-tuning
- it needs some supervision in the making
- the process is run frame by frame
- it generates a funscript file and a project file that you can restart from if you close the app
Handjob / Blowjob scenes:
- input consists in drawing the box where the penis is located
- the system detects hands and mouth, then tracks their position if touching the penis box
- applies a dynamic weighting mechanism for hands and mouth to determine the target funscript position
Other positions:
- input consists in drawing boxes on the male and female bodies
- these body parts are tracked and their relative distance is registered, then normalized
The process cuts sections of data every 15 seconds, auto filters, normalizes and export/saves funscript file and project.
There are still many things to implement before making it fully enjoyable, and needs to be optimized performance wise.
All this to actually come up with the following request : does any one of you have a VR video he would like to see a part scripted so as to test and provide feedback ? Please do not expect a fully scripted vid, but rather just short section(s) of it at this stage.
Just really looking for feedback, it still needs a LOT of work but it is already 2k+ lines of code and I kinda feel alone facing a giant mountain to hikeā¦
I would definitely like designing and coding help, but I am still very shy to share my terrible and under documented code at this stage.