Here in this code Both video and audio steaming is implemented.
App.py holds the flask code for steaming and cemara.py contains
video capture method. download full code here
#APP.py
from flask import Flask, Response,render_template
import pyaudio
from camera import VideoCamera
import cv2
app = Flask(__name__)
FORMAT = pyaudio.paInt16
CHANNELS = 2
RATE = 44100
CHUNK = 1024
RECORD_SECONDS = 5
audio1 = pyaudio.PyAudio()
def genHeader(sampleRate, bitsPerSample, channels):
datasize = 2000*10**6
o = bytes("RIFF",'ascii') # (4byte) Marks file as RIFF
o += (datasize + 36).to_bytes(4,'little') # (4byte) File size in bytes excluding this and RIFF marker
o += bytes("WAVE",'ascii') # (4byte) File type
o += bytes("fmt ",'ascii') # (4byte) Format Chunk Marker
o += (16).to_bytes(4,'little') # (4byte) Length of above format data
o += (1).to_bytes(2,'little') # (2byte) Format type (1 - PCM)
o += (channels).to_bytes(2,'little') # (2byte)
o += (sampleRate).to_bytes(4,'little') # (4byte)
o += (sampleRate * channels * bitsPerSample // 8).to_bytes(4,'little') # (4byte)
o += (channels * bitsPerSample // 8).to_bytes(2,'little') # (2byte)
o += (bitsPerSample).to_bytes(2,'little') # (2byte)
o += bytes("data",'ascii') # (4byte) Data Chunk Marker
o += (datasize).to_bytes(4,'little') # (4byte) Data size in bytes
return o
@app.route('/audio')
def audio():
# start Recording
def sound():
CHUNK = 1024
sampleRate = 44100
bitsPerSample = 16
channels = 2
wav_header = genHeader(sampleRate, bitsPerSample, channels)
stream = audio1.open(format=FORMAT, channels=CHANNELS,
rate=RATE, input=True,input_device_index=1,
frames_per_buffer=CHUNK)
print("recording...")
#frames = []
first_run = True
while True:
if first_run:
data = wav_header + stream.read(CHUNK)
first_run = False
else:
data = stream.read(CHUNK)
yield(data)
return Response(sound())
def gen(camera):
while True:
try:
frame = camera.get_frame()
yield (b'--frame\r\n' b'Content-Type: image/jpeg\r\n\r\n' + frame + b'\r\n\r\n')
except:
frame = cv2.imread('loading.jpg')
ret, jpeg = cv2.imencode('.jpg', frame)
frame = jpeg.tobytes()
yield (b'--frame\r\n' b'Content-Type: image/jpeg\r\n\r\n' + frame + b'\r\n\r\n')
@app.route('/video_feed')
def video_feed():
return Response(gen(VideoCamera()),
mimetype='multipart/x-mixed-replace; boundary=frame')
@app.route('/')
def index():
"""Video streaming home page."""
return render_template('index.html')
if __name__ == "__main__":
app.run(host='0.0.0.0', debug=True, threaded=True,port=5000)
camera.py
import cv2
import imutils
import numpy as np
import time
ds_factor=0.6
#net = cv2.dnn.readNetFromCaffe('deploy.prototxt.txt', 'res10_300x300_ssd_iter_140000.caffemodel')
class VideoCamera(object):
def __init__(self):
self.video = cv2.VideoCapture(0)
def __del__(self):
self.video.release()
def get_frame(self):
success, image = self.video.read()
#time.sleep(2.0)
frame = image
frame = imutils.resize(frame, width=400)
(h, w) = frame.shape[:2]
ret, jpeg = cv2.imencode('.jpg', frame)
return jpeg.tobytes()
index.html
<!DOCTYPE html>
<html lang="en">
<head>
<meta charset="UTF-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<meta http-equiv="X-UA-Compatible" content="ie=edge">
<title>Document</title>
<style type="text/css">
img {
display: block;
margin-left: auto;
margin-right: auto;
width: 60%;
height:60%;
}
div{
display: block;
margin-left: auto;
margin-right: auto;
width: 60%;
height:60%;
}
</style>
</head>
<body>
<h1 align="center">Video Streaming Demonistration</h1>
<img id="bg" class="center" src="{{ url_for('video_feed') }}">
<div class="center">
<audio style="width: 60%;" controls>
<source src="{{ url_for('audio') }}" type="audio/x-wav;codec=pcm">
Your browser does not support the audio element.
</audio>
</div>
</body>
</html>
Output