-
Notifications
You must be signed in to change notification settings - Fork 2
/
FER_WebApp.py
410 lines (282 loc) · 13.6 KB
/
FER_WebApp.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
# Importing required libraries, obviously
import logging
import logging.handlers
import threading
from pathlib import Path
import streamlit as st
import cv2
import numpy as np
from keras.models import model_from_json
from keras.preprocessing.image import img_to_array
from streamlit_webrtc import VideoTransformerBase, webrtc_streamer
import av
from typing import Union
try:
from typing import Literal
except ImportError:
from typing_extensions import Literal # type: ignore
def load_model(path):
json_file = open(path + 'model.json', 'r')
loaded_model_json = json_file.read()
json_file.close()
model = model_from_json(loaded_model_json)
model.load_weights(path + "model.h5")
print("Loaded model from disk")
return model
# Loading pre-trained parameters for the cascade classifier
try:
face_classifier = cv2.CascadeClassifier(cv2.data.haarcascades + 'haarcascade_frontalface_default.xml') # Face Detection
path = "/content/drive/MyDrive/Capstone Project 5/"
classifier = load_model(path)
# classifier =load_model('model.h5') #Load model
emotion_labels = ['Angry','Disgust','Fear','Happy','Neutral', 'Sad', 'Surprise'] # Emotion that will be predicted
except Exception:
st.write("Error loading cascade classifiers")
class VideoTransformer(VideoTransformerBase):
def transform(self, frame):
label=[]
img = frame.to_ndarray(format="bgr24")
face_detect = cv2.CascadeClassifier(cv2.data.haarcascades +'haarcascade_frontalface_default.xml')
emotion_labels = ['Angry','Disgust','Fear','Happy','Neutral', 'Sad', 'Surprise']
gray = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY)
faces = face_detect.detectMultiScale(gray, 1.3,1)
for (x,y,w,h) in faces:
a=cv2.rectangle(img,(x,y),(x+w,y+h),(0,255,0),2)
roi_gray = gray[y:y+h,x:x+w]
roi_gray = cv2.resize(roi_gray,(48,48),interpolation=cv2.INTER_AREA) ##Face Cropping for prediction
roi = roi_gray.astype('float')/255.0
roi = img_to_array(roi)
roi = np.expand_dims(roi,axis=0) ## reshaping the cropped face image for prediction
prediction = classifier.predict(roi)[0] #Prediction
label=emotion_labels[prediction.argmax()]
label_position = (x,y)
b=cv2.putText(a,label,label_position,cv2.FONT_HERSHEY_SIMPLEX,1,(0,255,0),2)
return b
def face_detect():
class VideoTransformer(VideoTransformerBase):
frame_lock: threading.Lock # `transform()` is running in another thread, then a lock object is used here for thread-safety.
in_image: Union[np.ndarray, None]
out_image: Union[np.ndarray, None]
def __init__(self) -> None:
self.frame_lock = threading.Lock()
self.in_image = None
self.out_image = None
def transform(self, frame: av.VideoFrame) -> np.ndarray:
in_image = frame.to_ndarray(format="bgr24")
out_image = in_image[:, ::-1, :] # Simple flipping for example.
with self.frame_lock:
self.in_image = in_image
self.out_image = out_image
return in_image
ctx = webrtc_streamer(key="snapshot", video_processor_factory=VideoTransformer)
while ctx.video_transformer:
with ctx.video_transformer.frame_lock:
in_image = ctx.video_transformer.in_image
out_image = ctx.video_transformer.out_image
if in_image is not None :
gray = cv2.cvtColor(in_image, cv2.COLOR_BGR2GRAY)
faces = face_classifier.detectMultiScale(gray)
for (x,y,w,h) in faces:
a=cv2.rectangle(in_image,(x,y),(x+w,y+h),(0,255,0),2)
roi_gray = gray[y:y+h,x:x+w]
roi_gray = cv2.resize(roi_gray,(48,48),interpolation=cv2.INTER_AREA) ##Face Cropping for prediction
if np.sum([roi_gray])!=0:
roi = roi_gray.astype('float')/255.0
roi = img_to_array(roi)
roi = np.expand_dims(roi,axis=0) ## reshaping the cropped face image for prediction
prediction = classifier.predict(roi)[0] #Prediction
label=emotion_labels[prediction.argmax()]
label_position = (x,y)
b=cv2.putText(a,label,label_position,cv2.FONT_HERSHEY_SIMPLEX,1,(0,255,0),2) # Text Adding
st.image(b,channels="BGR")
from streamlit_webrtc import (
ClientSettings,
VideoTransformerBase,
WebRtcMode,
webrtc_streamer,
)
HERE = Path(__file__).parent
logger = logging.getLogger(__name__)
WEBRTC_CLIENT_SETTINGS = ClientSettings(
rtc_configuration={"iceServers": [{"urls": ["stun:stun.l.google.com:19302"]}]},
media_stream_constraints={"video": True, "audio": True},
)
def about():
st.write(
'''
**Haar Cascade** is an object detection algorithm.
It can be used to detect objects in images or videos.
The algorithm has four stages:
1. Haar Feature Selection
2. Creating Integral Images
3. Adaboost Training
4. Cascading Classifiers
Read more :
point_right:
https://docs.opencv.org/2.4/modules/objdetect/doc/cascade_classification.html
https://sites.google.com/site/5kk73gpu2012/assignment/viola-jones-face-detection#TOC-Image-Pyramid
''')
def app_video_filters():
""" Video transforms with OpenCV """
class OpenCVVideoTransformer(VideoTransformerBase):
type: Literal["noop", "cartoon", "edges", "rotate"]
def __init__(self) -> None:
self.type = "noop"
def transform(self, frame: av.VideoFrame) -> av.VideoFrame:
img = frame.to_ndarray(format="bgr24")
if self.type == "noop":
pass
elif self.type == "cartoon":
# prepare color
img_color = cv2.pyrDown(cv2.pyrDown(img))
for _ in range(6):
img_color = cv2.bilateralFilter(img_color, 9, 9, 7)
img_color = cv2.pyrUp(cv2.pyrUp(img_color))
# prepare edges
img_edges = cv2.cvtColor(img, cv2.COLOR_RGB2GRAY)
img_edges = cv2.adaptiveThreshold(
cv2.medianBlur(img_edges, 7),
255,
cv2.ADAPTIVE_THRESH_MEAN_C,
cv2.THRESH_BINARY,
9,
2,
)
img_edges = cv2.cvtColor(img_edges, cv2.COLOR_GRAY2RGB)
# combine color and edges
img = cv2.bitwise_and(img_color, img_edges)
elif self.type == "edges":
# perform edge detection
img = cv2.cvtColor(cv2.Canny(img, 100, 200), cv2.COLOR_GRAY2BGR)
elif self.type == "rotate":
# rotate image
rows, cols, _ = img.shape
M = cv2.getRotationMatrix2D((cols / 2, rows / 2), frame.time * 45, 1)
img = cv2.warpAffine(img, M, (cols, rows))
return img
webrtc_ctx = webrtc_streamer(
key="opencv-filter",
mode=WebRtcMode.SENDRECV,
client_settings=WEBRTC_CLIENT_SETTINGS,
video_transformer_factory=OpenCVVideoTransformer,
async_transform=True,
)
if webrtc_ctx.video_transformer:
webrtc_ctx.video_transformer.type = st.radio(
"Select transform type", ("noop", "cartoon", "edges", "rotate")
)
st.markdown(
"This demo is based on "
"https://github.com/aiortc/aiortc/blob/2362e6d1f0c730a0f8c387bbea76546775ad2fe8/examples/server/server.py#L34. " # noqa: E501
"Many thanks to the project."
)
def main():
activities = ["Introduction","Home","Real-Time Snapshot", "Check Camera","About","Contact Us"]
choice = st.sidebar.selectbox("Pick something Useful", activities)
if choice == "Real-Time Snapshot":
html_temp = """
<body style="background-color:red;">
<div style="background-color:teal ;padding:10px">
<h2 style="color:white;text-align:center;">Face Emotion Recognition WebApp</h2>
</div>
</body>
"""
st.markdown(html_temp, unsafe_allow_html=True)
st.write("**Using the Haar cascade Classifiers**")
st.write("Go to the About section from the sidebar to learn more about it.")
st.write("**Instructions while using the APP**")
st.write('''
1. Click on the Start button to start.
2. WebCam window will open automatically.
3. It will automatically throw the image with the prediction at that instant.
4. Make sure that camera shouldn't be used by any other app.
5. For live recognition the app is getting slow and takes more time to predict and couldn't predict easily thus fluctuating the result.
Thus Taking a snapshot at any instant of time and it will automatically predict and give the picture with prediction.
6. Easy to know what was or what is the emotion at a particular time.
7. Click on Stop to end.
8. Still webcam window didnot open, go to Check Camera from the sidebar.''')
face_detect()
elif choice =="Home":
html_temp = """
<body style="background-color:red;">
<div style="background-color:teal ;padding:10px">
<h2 style="color:white;text-align:center;">Face Emotion Recognition WebApp</h2>
</div>
</body>
"""
st.markdown(html_temp, unsafe_allow_html=True)
st.write("**Using the Haar cascade Classifiers**")
st.write("Go to the About section from the sidebar to learn more about it.")
st.write("**Instructions while using the APP**")
st.write('''
1. Click on the Start button to start.
2. WebCam window will open automatically.
3. It will automatically predict at that instant.
4. Make sure that camera shouldn't be used by any other app.
5. Click on Stop to end.
6. Still webcam window didnot open, go to Check Camera from the sidebar.
''')
webrtc_streamer(key="example", video_transformer_factory=VideoTransformer)
elif choice == "Check Camera":
html_temp = """
<body style="background-color:red;">
<div style="background-color:teal ;padding:10px">
<h2 style="color:white;text-align:center;">Check Webcam is working or not</h2>
</div>
</body>
"""
st.markdown(html_temp, unsafe_allow_html=True)
st.write("**Instructions while Checking Camrea**")
st.write('''
1. Click on Start to open webcam.
2. If you have more than one camera , then select by using select device.
3. Have some fun with your camera by choosing the options below.
4. Click on Stop to end.
5. Still webcam window didnot open, Contact Us.''')
app_video_filters()
elif choice == "About":
html_temp = """
<body style="background-color:red;">
<div style="background-color:teal ;padding:10px">
<h2 style="color:white;text-align:center;">Haar Cascade Object Detection</h2>
</div>
</body>
"""
st.markdown(html_temp, unsafe_allow_html=True)
about()
elif choice=="Contact Us":
with st.form(key='my_form'):
text_input = st.text_input(label='Enter some text')
submit_button = st.form_submit_button(label='Submit')
st.write('''
Email:- 1) [email protected].
Linkedin:-https://www.linkedin.com/in/sajal-sinha/
''')
html_temp = """
<body style="background-color:white;">
<div style="background-color:red ;padding:0.25px">
<h4 style="color:white;text-align:center;">Copyright © 2021 | Sajal Atif </h4>
</div>
</body>
"""
st.markdown(html_temp, unsafe_allow_html=True)
elif choice=="Introduction":
html_temp = """
<body style="background-color:white;">
<h1 style="text-align:center;"> Face Emotion Recognitiom </h1>
<h3 style="color:black;text-align:center;">During online classes students often tends to loose attention, which leads to
overall non-productivity. For a teacher, its often important for its students to easily
grasp concept taught by them. Teachers have skills to observe their students and im-
prove their way throughout their teaching. But due to online teaching, observing has
become tough which has eventually disturbed student teacher balance and teaching
methods. So, our aim was to develop a Face-Emotion-Recognition Model which can be
used a micro service as well so that teachers can understand students much better and
enlighten the way to teach.</h3>
<h3 style="color:red;text-align:center;">To Know your emotion proceed to Home from the side bar.</h3>
</div>
</body>
"""
st.markdown(html_temp, unsafe_allow_html=True)
if __name__ == "__main__":
main()