Projects

Context-Aware Zero-Shot Anomaly Detection in Surveillance Using Contrastive and Predictive Spatiotemporal Modeling
CSE 400 — Final Year Thesis

Context-Aware Zero-Shot Anomaly Detection in Surveillance Using Contrastive and Predictive Spatiotemporal Modeling

Supervisors: Prof. Dr. Md. Ashraful Alam, Md Tanzim Reza

Zero-Shot LearningVideo Anomaly DetectionComputer VisionContrastive LearningTransformer Architecture

Developed a novel zero-shot anomaly detection framework that identifies abnormal events in surveillance footage without requiring any anomaly examples during training. The system combines spatiotemporal transformers with vision-language understanding to detect previously unseen threats in real-time.

Key Innovation: Introduced a dual-stream architecture integrating TimeSformer for spatiotemporal feature extraction, DPC-RNN for predictive temporal modeling, and CLIP for semantic context alignment—enabling true zero-shot detection with context awareness.

Technical Approach: Jointly trained the model using InfoNCE and Contrastive Predictive Coding (CPC) losses to learn normal behavior patterns. A context-gating mechanism modulates predictions based on scene-specific cues, reducing false alarms by adapting to different surveillance environments.

Results: Achieved 84.5% ROC-AUC and 72.3% PR-AUC on the UCF-Crime dataset, outperforming state-of-the-art zero-shot methods including AnomalyCLIP (82.4%) while maintaining competitive detection latency of 0.45 seconds.

Tech Stack: Python, PyTorch, Hugging Face Transformers, TimeSformer, CLIP, DPC-RNN, OpenCV, UCF-Crime dataset

Blue-Light Blocking Glasses Using Machine Learning
CSE 424 — Pattern Recognition

Blue-Light Blocking Glasses Using Machine Learning

Instructors: Annajiat Alim Rasel, Sadiul Arefin Rafi

Machine LearningRegression ModelsSpectrophotometryHealth InformaticsData Analysis

Developed machine learning models to predict and analyze the spectrophotometric properties of commercial blue-light blocking lenses, evaluating their effectiveness in filtering circadian-proficient wavelengths (455-560nm) while preserving visual clarity across diverse lighting conditions.

Research Methodology: Tested 50 commercial blue-blocking lenses under 5 distinct light sources (sunlight, fluorescent, incandescent, LED, and tablet displays), measuring absolute irradiance across the visible spectrum (380-780nm).

Model Performance: Implemented KNN, SVM, and Linear Regression. The KNN model achieved superior performance with 91.4% R² score, 5.921 RMSE, and 3.844 MAE, significantly outperforming SVM (48.3% R²) and Linear Regression (79.7% R²).

Key Findings: Red-tinted lenses transmitted the least circadian-efficient light, while orange-tinted glasses demonstrated the highest transmission specificity for blocking physiologically active light in daylight conditions.

Tech Stack: Python, Scikit-learn, Pandas, NumPy, Matplotlib, OpenRefine, SVM, KNN, Linear Regression, Spectrophotometric Data Analysis

Comparative Evaluation of Clustering Algorithms on the Wine Dataset with Stability Analysis
CSE 425 — Neural Networks

Comparative Evaluation of Clustering Algorithms on the Wine Dataset with Stability Analysis

Instructor: Moin Mostakim

Unsupervised LearningVariational AutoencodersClustering AnalysisUncertainty QuantificationNeural Networks

Developed a novel Stochastic Clustering Neural Network (SCNN) inspired by Variational Autoencoders to perform uncertainty-aware clustering on the Wine dataset, incorporating probabilistic latent representations to quantify cluster assignment confidence.

Novel Architecture: Designed a VAE-inspired stochastic neural network featuring an encoder-decoder structure with reparameterization trick, enabling gradient-based optimization through stochastic sampling while learning meaningful latent representations.

Comparative Analysis: Conducted comprehensive evaluation against classical baselines (K-Means, GMM, SOM) using Silhouette Score, Adjusted Rand Index (ARI), and Normalized Mutual Information (NMI). Introduced stability analysis across multiple random seeds.

Key Findings: While GMM achieved highest accuracy (78.44% ARI), the SCNN provided unique uncertainty quantification capabilities with 0.4885 stability variance. SOM demonstrated extreme instability (974.89), highlighting sensitivity to initialization.

Tech Stack: Python, PyTorch, Scikit-learn, NumPy, Pandas, Matplotlib, Seaborn, Wine Dataset

Loan Eligibility Prediction Using Machine Learning Models with SMOTE for Class Imbalance
CSE 427 — Machine Learning

Loan Eligibility Prediction Using Machine Learning Models with SMOTE for Class Imbalance

Instructor: Prof. Dr. Chowdhury Mofizur Rahman

Machine LearningEnsemble MethodsClass ImbalanceSMOTEFinancial ML

Built an automated loan eligibility prediction system for Dream Housing Finance Company using machine learning classifiers to streamline the approval process based on applicant demographics, financial data, and credit history.

Data Preprocessing: Implemented comprehensive preprocessing including median/mode imputation for missing values, one-hot encoding, feature scaling, and SMOTE to address severe class imbalance in loan approval data.

Model Comparison: Evaluated five ML algorithms—Random Forest, Logistic Regression, AdaBoost, KNN, and MLP—using accuracy, precision, recall, and F1-score metrics. Random Forest and MLP achieved highest accuracy (89.29%), with MLP demonstrating exceptional recall (96%).

Key Insights: Correlation analysis revealed ApplicantIncome (0.57 with LoanAmount), CreditHistory, and LoanAmount as strongest predictors. However, all models struggled with False class prediction, indicating need for advanced imbalance handling.

Tech Stack: Python, Scikit-learn, Pandas, NumPy, Matplotlib, Seaborn, Imbalanced-learn (SMOTE), ensemble methods

Connect

mohammodtareqaziz@gmail.comMerul Badda, Dhaka-1212, Bangladesh+88 01922-667040