Focusing on Convolutional Neural Network (CNN), ResNet, DenseNet, and Vision Transformer (ViT) models involving training on feature-augmented image data to capture complex visual characteristics. This was implemented using Python programming language on Google Colab.
Developed a generative model using pre-trained models capable of producing realistic, high-quality synthetic faces indistinguishable from real faces. This was implemented using Python programming language on Google Colab.
Trained and improved the ability of a Convolutional Neural Network to identify if a patient has pneumonia using X-ray images. Created a local website to display the prediction when a new X-ray image is uploaded. Training of the model was done using Python programming language and deployment of the model to a local webpage was done using Streamlit.
Objective: The goal of this Mini-Hackathon was to predict whether the literacy rate was high/ medium/ low in the different districts of India.
Approach: Extensive data-pre-processing was required to clean the data, remove missing data and values, transform the data by normalization, select uncorrelated features and integrate the data into a single dataframe. After the data was pre-processed, a machine learning classifier was trained to obtain a model that could accurately predict the literacy rate (high/ medium/ low).
Objective: The goal of this Mini-Hackathon was to identify the unique writing styles of different authors using a machine learning based approach.
Approach: A manual, stylometric analysis using hand-crafted features was also used for comparison. A machine learning model was trained on textual data with corresponding author. Bag of Words and Word2Vec based approaches were used for feature representation and analysis.
Objective: The goal of this Mini-Hackathon was to classify fashion clothing by building a machine learning model and improving its performance in recognizing fashion clothing images.
Approach: Firstly, features were extracted from the raw images or PCA applied images. A multi-layer perceptron was trained on the features and model performance was evaluated. The model performance was improved using different hyperparameter values
Objective: The goal of this Mini-Hackathon was to obtain the sales forecast for a big retail store. This was presented to us as a Kaggle competition.
Approach: The data consisted of a real-world dataset that included the monthly sales. Extensive data-preprocessing was required to clean the data and integrate the data into a single data frame. We completed the competition successfully by building a time-series based machine learning model using linear regression that accurately forecasted the monthly and product wise sales for the retail store.
Objective: The objective of this Mini-Hackathon was to identify and classify simple vocal utterances such as yes or no.
Approach: >2600 audio samples were used to extract useful feature representations in the form of Mel Frequency Cepstral Coefficients (MFCC). A machine learning classifier was trained on the MFCC features and the model was accurately able to classify the test audio samples as yes or no
Objective: The goal of this hackathon was to create and deploy a food ordering application, using speech recognition AI techniques.
Approach: Training a model using features extracted from voice samples and to integrate a machine learning classifier that could identify the correct food items and servings. The final user interface accurately identified the food item from a voice command and confirmed the order as a response.
Objective: The goal of this hackathon was to build a conversational bot to interact with the user and achieve the desired outcome
Approach: Two approaches were used in achieving the task - (i) Amazon Alexa API based bot and (ii) Python coding-based bot. Python coding-based bot has to identify the statement given by the end-user and match the requirement with any previous training examples and give an output appropriately.
Objective: The objective of this Mini-Hackathon was to classify images of dogs and cats
Approach: A Pytorch framework was used to build and train a Convolutional Neural Network (CNN) model on a dataset of >22000 images of cats and dogs. Importantly, image classification was achieved with the CNN model and an accuracy of >90% was obtained.
Objective: In this Hackathon, we implemented AI based Face and Expression recognition to build an antiface spoofing application. The goal was to unlock a mobile application based on the accurate identification of a user’s face and expression, as a safety feature.
Approach: Firstly, a Siamese Network based representation was obtained from a large dataset of face images and used for detecting Face Similarity. Then we built a small dataset of our team members’ face images and expressions and applied a deep learning-based CNN model to perform Face and Expression Recognition. It was an amazing experience to unlock the mobile app with our Face and Expression Recognition model, at the end of the project and course.
Description: A project based on GPS and Arduino to measure the distance between two points using the Haversine formula, which calculates the distance between two points on a sphere from their longitudes and latitudes. This uses a GPS system and keeps a record of the latitude and logitude of the starting point and tracks them till the end button is clicked. The latitude and logitude os the start and end points are then given to the arduino as input. Using these values, arduino calculates the distance between them and displays it using an digital display. This was further improvised to calculate distance instead of displacement.
Description: A project using Arduino and 3 Easy Drivers. It utilizes the INKSCAPE software to draw or write within a specified space, with movement managed by stepper motors. The inkscape takes the structure of we want to draw and creates a list of movements needed in length, width and height. These list of movements are the input to the arduino which further makes decisions on which easy drive to move. The end of the drive is a pen like structure which captures the image in a paper.
Description: A device using Raspberry Pi to monitor health status of senior citizens and alert emergency contacts in case of abnormalities. We continuously track the heart rate, temperature, O2 level etc. using sensors and refresh the information for every 10 seconds. If any of these values are below the desired values, then this device initiates a alarm. This alarm sends a text message to the emergency contacts and alerts the surroundings.
A model which is used to predict which bin a particular product is stowed into in a pod using videos at the Amazon Fulfillment Centre.
A model which is used to identify human bodies in images and determine if the social distancing principles were followed at the Amazon Fulfillment Centres.
A model used to classify whether a package in an image is damaged or not and the type of damage.
A model used to check the compliance and safety of the pods and trays in the trucks using the images.
A model which checks whether a product is within or outside the package using images.
A model used to identify movable objects in an image which is used by small size robots for moving directions.
A model used to classify whether a robotic arm is able to pick and place an item correctly using video based data.
Optimizing a deep learning model for Cost Per Click/Impression Prediction.
Automated preprocessing workflows in Dataiku, reducing data preparation time and improving data quality.
© Lakshmi Keerthi Gopireddy