Splitting up and fake webcam
This commit is contained in:
parent
c3f2738e78
commit
97a758bdb3
3 changed files with 110 additions and 32 deletions
41
camera.py
Normal file
41
camera.py
Normal file
|
@ -0,0 +1,41 @@
|
||||||
|
from threading import Thread, Event, Lock
|
||||||
|
import cv2
|
||||||
|
|
||||||
|
class Camera(Thread):
|
||||||
|
def __init__(self):
|
||||||
|
Thread.__init__(self, daemon=True)
|
||||||
|
self.finished = Event()
|
||||||
|
self.ready = Event()
|
||||||
|
self._cap = cv2.VideoCapture(0)
|
||||||
|
self._frame = None
|
||||||
|
self.frame_lock = Lock()
|
||||||
|
|
||||||
|
def stop(self):
|
||||||
|
"""Stop the timer if it hasn't finished yet."""
|
||||||
|
self.finished.set()
|
||||||
|
|
||||||
|
def read(self):
|
||||||
|
f = None
|
||||||
|
self.ready.wait()
|
||||||
|
f = self.frame
|
||||||
|
self.ready.clear()
|
||||||
|
return f
|
||||||
|
|
||||||
|
@property
|
||||||
|
def frame(self):
|
||||||
|
f = None
|
||||||
|
with self.frame_lock:
|
||||||
|
f = self._frame
|
||||||
|
return f
|
||||||
|
|
||||||
|
@frame.setter
|
||||||
|
def frame(self, f):
|
||||||
|
with self.frame_lock:
|
||||||
|
self._frame = f
|
||||||
|
|
||||||
|
def run(self):
|
||||||
|
while not self.finished.is_set():
|
||||||
|
_, self.frame = self._cap.read()
|
||||||
|
self.ready.set()
|
||||||
|
self._cap.release()
|
||||||
|
|
98
main.py
98
main.py
|
@ -1,42 +1,78 @@
|
||||||
import zmq
|
import zmq
|
||||||
import cv2
|
import cv2
|
||||||
|
from pyfakewebcam import FakeWebcam
|
||||||
import numpy as np
|
import numpy as np
|
||||||
|
from camera import Camera
|
||||||
|
from signal import signal, SIGINT
|
||||||
|
|
||||||
ctx = zmq.Context()
|
def process_mask(m, shape):
|
||||||
sock = ctx.socket(zmq.REQ)
|
w, h = shape
|
||||||
sock.connect('ipc:///tmp/bodypix')
|
m = np.frombuffer(m, dtype=np.uint8).reshape((w, h, -1))
|
||||||
cap = cv2.VideoCapture(0)
|
m = m[:, :, 0].astype(np.float32)
|
||||||
while True:
|
m = cv2.blur(cv2.UMat(m), (30, 30))
|
||||||
_, frame = cap.read()
|
return m.get()
|
||||||
_, image = cv2.imencode('.jpg', frame)
|
|
||||||
|
|
||||||
print("LENGTH", len(image.tostring()), flush=True)
|
def composite(foreground, backdrop, fore_mask):
|
||||||
sock.send(image.tostring())
|
inv_mask = 1 - fore_mask
|
||||||
convereted_img = sock.recv()
|
for c in range(foreground.shape[2]):
|
||||||
|
foreground[:, :, c] = foreground[:, :, c] * fore_mask + backdrop[:, :, c] * inv_mask
|
||||||
|
return foreground
|
||||||
|
|
||||||
mask = np.frombuffer(convereted_img, dtype=np.uint8)
|
|
||||||
mask = mask.reshape((frame.shape[0], frame.shape[1], 4))
|
|
||||||
mask = mask[:, :, 0]
|
|
||||||
|
|
||||||
# post-process mask and frame
|
def get_background(uri=None):
|
||||||
mask = cv2.UMat(mask)
|
"""
|
||||||
mask = cv2.dilate(mask, np.ones((10, 10), np.uint8), iterations=1)
|
Either grabs the image specified or
|
||||||
mask = cv2.blur(cv2.UMat(mask.get().astype(np.float32)), (30, 30))
|
blurs the current background.
|
||||||
|
"""
|
||||||
|
initial_frame = camera.read()
|
||||||
|
w, h = initial_frame.shape[0], initial_frame.shape[1]
|
||||||
|
backdrop = None
|
||||||
|
if uri is not None:
|
||||||
|
backdrop = cv2.resize(
|
||||||
|
cv2.UMat(cv2.imread('/home/rozek/Pictures/IMG_20191013_181848.jpg')),
|
||||||
|
(h, w)
|
||||||
|
).get().astype(np.uint8)
|
||||||
|
else:
|
||||||
|
backdrop = cv2.GaussianBlur(initial_frame, (221, 221), sigmaX=20, sigmaY=20).astype(np.uint8)
|
||||||
|
return backdrop, (w, h)
|
||||||
|
|
||||||
frame = cv2.UMat(frame)
|
|
||||||
background = cv2.GaussianBlur(frame, (221, 221), sigmaX=20, sigmaY=20)
|
|
||||||
|
|
||||||
# composite the foreground and background
|
def stop(signal_received, stack_frame):
|
||||||
frame = frame.get().astype(np.uint8)
|
"""Gracefully stops the application."""
|
||||||
mask = mask.get().astype(np.float32)
|
global running
|
||||||
background = background.get().astype(np.uint8)
|
running = False
|
||||||
inv_mask = 1 - mask
|
print("Stopping...")
|
||||||
for c in range(frame.shape[2]):
|
|
||||||
frame[:, :, c] = frame[:, :, c] * mask + background[:, :, c] * inv_mask
|
|
||||||
|
|
||||||
cv2.imshow('frame', frame)
|
|
||||||
if cv2.waitKey(1) & 0xFF == ord('q'):
|
|
||||||
break
|
|
||||||
|
|
||||||
cap.release()
|
# Start Camera
|
||||||
cv2.destroyAllWindows()
|
camera = Camera()
|
||||||
|
camera.start()
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
signal(SIGINT, stop)
|
||||||
|
|
||||||
|
# Setup ZeroMQ
|
||||||
|
ctx = zmq.Context()
|
||||||
|
sock = ctx.socket(zmq.REQ)
|
||||||
|
sock.connect('ipc:///tmp/bodypix')
|
||||||
|
|
||||||
|
background, (width, height) = get_background()
|
||||||
|
fake = FakeWebcam('/dev/video20', height, width)
|
||||||
|
|
||||||
|
running = True
|
||||||
|
print("Running...")
|
||||||
|
|
||||||
|
while running:
|
||||||
|
frame = camera.read()
|
||||||
|
_, image = cv2.imencode('.jpg', frame)
|
||||||
|
frame = frame.astype(np.uint8)
|
||||||
|
|
||||||
|
# Process image to find body masks
|
||||||
|
sock.send(image.tostring())
|
||||||
|
mask = process_mask(sock.recv(), (width, height))
|
||||||
|
|
||||||
|
frame = composite(frame, background, mask)
|
||||||
|
|
||||||
|
# Send to the fake camera device
|
||||||
|
frame = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB)
|
||||||
|
fake.schedule_frame(frame)
|
||||||
|
|
|
@ -25,6 +25,7 @@ async function run() {
|
||||||
const image = decodeJpeg(msg)
|
const image = decodeJpeg(msg)
|
||||||
const segmentation = await net.segmentPerson(image);
|
const segmentation = await net.segmentPerson(image);
|
||||||
await sock.send(segmentation.data);
|
await sock.send(segmentation.data);
|
||||||
|
tf.dispose(image)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue