Deep Learning Projects
Relevant Coursework
CS 6746: Computer Vision
CSE 6740: Computational Data Analysis (ML)
CSE 8803: Data Science in Social Networks
CS 7650: Natural Language Processing
Relevant Coursework
CS 6746: Computer Vision
CSE 6740: Computational Data Analysis (ML)
CSE 8803: Data Science in Social Networks
CS 7650: Natural Language Processing
Bridged mutism communication gap with computer vision and deep learning. Developed an AI tool that converts
video livestream to textual and audio output.
Used OpenCV and ffmpeg algorithms to track hand landhand sign landmarks in real-time.
Then developed a CNN based attention mechanism model with TensorFlow to recognize hand sign alphabets at 90% accuracy.
Finetuned Huggingface GPT-neo-x model with 20B parameters with QLoRA,
reducing trainable parameters to 0.08% in Colab environment to enable faster training.
Deployed RAG with cosine-similarity retrieval to enable output optimization and reduce hallucination.
Applies Bayesian optimization to workwith expensive-to-evaluate LLMs.
Bocoel uses Bayesian optimization as a backbone for exploring the embedding space given by our corpus, which allows it to select a good subset acting as a mini snapshot of the corpus.
Bocoel works in the following steps:
1) Encode individual entry into embeddings (cheaper & faster than LLM and reusable).
2) Use Bayesian optimization to select queries to evaluate.
3) Use the queries to retrieve from our corpus (with the encoded embeddings).
4) The evaluations generated are easily managed by the provided manager utility.
To our knowledge, this is the first work aiming to reduce computation costs during evaluation (benchmarking) with a (possibly dynamic) budget.
Developed a BERT embedded tripartite Graph Convolutional Network (GCN) to accurately
predicted Amazon customer dissatisfaction score .
By leveraging GCN on tripartite graph using BERT embeddings of 30+gb amazon user-product-eview text dataset as nodes in PyTorch,
reached at near-perfect 0.98 R2 regression score.
During training, boosted accuracy over 7% by applying multitask learning on human-annotated review helpfulness and sentiment labels to beat multimodal (BERT+CLIP) baseline by significant margins.