Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Vedanth futurex #5

Open
wants to merge 10 commits into
base: main
Choose a base branch
from
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
135 changes: 8 additions & 127 deletions README.md
Original file line number Diff line number Diff line change
@@ -1,132 +1,13 @@
# FER May Hakathon
## Emotion Detection FER Hackathon - Vedanth

Facial Emotion Detection Hackathon Project, Create a model and test it uing 5 to 10 sec videos to detect emotions
Team Name : FutureX
Team Captain : Vedanth Aggarwal
Members : Vedanth Aggarwal

# Facial Emotion Recognition
Model Building : Advanced CNN architectures + Tranfer Learning

<div id="top"></div>
<div align="center">
Deployment : Streamlit + OpenCV

![Python](https://img.shields.io/badge/python-3670A0?style=for-the-badge&logo=python&logoColor=ffdd54)
![TensorFlow](https://img.shields.io/badge/TensorFlow-%23FF6F00.svg?style=for-the-badge&logo=TensorFlow&logoColor=white)
![Jupyter Notebook](https://img.shields.io/badge/jupyter-%23FA0F00.svg?style=for-the-badge&logo=jupyter&logoColor=white)
The videos are the files with names like recording, angry, happy_multiface


</div>

<br />
<div align="center">
<img src="assets/falcons-logo2.png" alt="Logo" >
</div>
<br /><br />
<div align="center">
<img src="assets/dds_logo.png" alt="DDS logo" >
</div>
# Decoding Data Science in partnership with Falcons.ai
<br /><br />

Objective: Develop an efficient facial emotion classification system employing OpenCV/Tensorflow to identify facial emotions within video streams. The goal is to achieve a high level of accuracy, low latency, and minimal computational overhead.

Similar to: <br/>

[![IMAGE ALT TEXT HERE](https://img.youtube.com/vi/Msk1drgWEdY/0.jpg)](https://www.youtube.com/watch?v=Msk1drgWEdY)

Data Source: A video dataset or a combination of image datasets featuring the target objects in states of emotion.

Kaggle : https://www.kaggle.com/datasets/msambare/fer2013

Preprocessing (if needed): Standardize or augment the images/video frames to improve model generalization, if necessary, while preserving the aspect ratio and critical features.

Model Selection & Training:
1. Using the FER dataset(partial).
2. Train a custom model using the prepared dataset and analyze the performance.
3. Deploy Streamlit and OpenCV to allow users a web ui in which to upload a video and have the video frames analyzed by your model.

Expecation

The expectations are for the following:
1) The code used to train the model.
2) The model you trained.
3) The Code used to run the UI and upload the video for inference.

This problem set provides a clear path to address image analysis issues using OpenCV, with a focus on Facial Emotion Classification in video streams. It allows researchers or students to hone in on critical aspects such as data preprocessing, model selection, hyperparameter tuning, performance evaluation, and results interpretation.
<br /><br />

-------------- Fully functional Jupyternotebook will be added upon hack-a-thon challenge completion --------------


</p>
<p align="right">(<a href="#top">back to top</a>)</p>
<br />

<!-- How to use -->
## Usage
<br />
<p>
To use the notebook with relative ease please follow the steps below:
<br />
</p>

1. Ensure all of the required libraries are installed.

2. Load the libraries.

3. Run the cells and the cloud images will be generated and saved in the "clouds" directory.

</p>
<br />
<p align="right">(<a href="#top">back to top</a>)</p>
<br />




<!-- CONTRIBUTING -->
## Contributing

Contributions are what make the open source community such an amazing place to learn, inspire, and create. Any contributions you make are **greatly appreciated**.

If you want, feel free to fork this repository. You can also simply open an issue with the tag "enhancement".
Don't forget to give the project a star! Thanks again!

1. Fork the Project
2. Create your Feature Branch (`git checkout -b feature/YourFeature`)
3. Commit your Changes (`git commit -m 'Add some YourFeature'`)
4. Push to the Branch (`git push origin feature/YourFeature`)
5. Open a Pull Request
<br />


<p align="right">(<a href="#top">back to top</a>)</p>



<!-- LICENSE -->
## License

![](https://img.shields.io/badge/License-MIT-blue)

<p align="right">(<a href="#top">back to top</a>)</p>



<!-- CONTACT -->
## Contact

Project Link: [https://github.com/Falcons-ai/fer_dds_challenge]


<p align="right">(<a href="#top">back to top</a>)</p>



Contributing
Contributions are what make the open source community such an amazing place to learn, inspire, and create. Any contributions you make are greatly appreciated.

If you want, feel free to fork this repository. You can also simply open an issue with the tag "enhancement". Don't forget to give the project a star! Thanks again!

Fork the Project
Create your Feature Branch (git checkout -b feature/YourFeature)
Commit your Changes (git commit -m 'Add some YourFeature')
Push to the Branch (git push origin feature/YourFeature)
Open a Pull Request
*Project works with multiple different people in a video ( can detect many face emotions simultaneously )
Binary file added Recording 2024-06-04 104505.mp4
Binary file not shown.
Binary file added Recording 2024-06-04 104648.mp4
Binary file not shown.
Binary file added angry.mp4
Binary file not shown.
Loading