Hi, I'm Mohamad Zamini.

A
Self-driven and passionate about advancing machine learning, I am focused on pushing the boundaries of Large Language Models (LLMs) through innovative research and development. As a final-year PhD Candidate with hands-on experience in optimizing LLMs during my recent internship, I am committed to solving complex real-world problems with cutting-edge AI technology. My ambition is to contribute to the future of AI by developing scalable and efficient models that can transform industries and enhance human-computer interaction.

About

In advancing Visual Question Answering (VQA) models, I utilized innovative methodologies. This involved generating a new dataset and designing interpretable VQA models specifically for the Semi-CLEVR dataset. Leveraging large language models (LLMs), causal explanations were developed to enhance model interpretability. Additionally, diffusion models were applied to the novel CLEVR dataset to accurately decompose occluded shapes, capitalizing on their capacity to model complex data distributions and uncover hidden details. This research aimed to enhance generalization in VQA through the integration of causal reasoning and advanced modeling techniques.

  • Programming Languages: Python, C++
  • Databases: MySQL, MongoDB, PostgreSQL
  • Libraries & Frameworks: PyTorch, TensorFlow, Hugging Face Transformers, Keras, NumPy, Pandas, OpenCV
  • Model Optimization & Deployment: ONNX, TensorRT, TorchServe, FastAPI
  • Tools & Platforms: Git, Docker, Kubernetes, AWS, GCP, Azure, JIRA, Weights & Biases (wandb)

Seeking a challenging position that leverages my expertise in Machine Learning and Software Engineering, offering opportunities for professional development, innovative experiences, and personal growth.

Experience

Machine Learning Engineer Intern
  • Fine-tuned LLM models, including Mistral, LLaMA, and GPT, leveraging techniques such as activation sparsity and attention sparsity to optimize performance.
  • Applied techniques such as KWTA, dynamic context pruning, and KV caching to enhance model efficiency.
  • Tools: Python, PyTorch, Accelerate, GPT, llama
July 2024 - Sept 2024 | Redwood city, CA
Digital Innovation Intern
  • During my internship at Petrolern as a Digital Innovation Intern, I gained experience in both machine learning and data compression techniques
  • I developed a semantic compression technique using a deep autoencoder to effectively map data tuples into a lower-dimensional representation
  • As a machine learning engineer, I built models for analyzing geothermal data and improved their performance through algorithmic optimization
June 2022 - Aug 2022 | Atlanta, GE
NLP Engineer
  • Fine-tuned models like BART for summarization on Persian text data.
  • Implemented Matrix Factorization for topic modeling.
  • utilized BiLSTM-CRF Models for sequential tagging.
  • Tools: Python, Scikit-learn, NLTK
June 2018 - Aug 2019 | Tehran, Iran

Projects

music streaming app
Explainability analysis

SHAP vs Lime Vs ELI5

Accomplishments
  • Tools: Python, PyTorch
  • To explain the model's predictions, the project uses model interpretability tools such as SHAP (SHapley Additive exPlanations), Lime (Local Interpretable Model-agnostic Explanations), and Eli5 (Explain Like I'm 5). These tools provide insights into how the model makes decisions and highlight the importance of different features in predicting strokes..
Causal Inference
Causal Inference

Causal Inference with Bayesian Networks

Accomplishments
  • Tools: Python, PyTorch
Screenshot of web app
Bi-directional Autoregressive Transformers From scratch

A simple Bi-directional Autoregressive Transformers From scratch.

Accomplishments
  • Tools: Python, PyTorch
  • implement the tokenizer, the model and tune it just the way we want.
Screenshot of  web app
image captioning

VIT + GPT2 image captioning

Accomplishments
  • Incorporated Convolution Neural Networks (CNN) for extracting image features and Long Short Term Memory for extracting question embeddings.
Screenshot of  web app
Attention-based Graph Neural Network

Multi-Label Text Classification using Attention-based Graph Neural Network.

Accomplishments
  • Multi-Label-Text-Classification-using-Attention-based-Graph-Neural-Network.
Screenshot of  web app
GPT2 for writing Python code

explore how to finetune the GPT2 and create a Python Question answering mdoel like chatgpt

Accomplishments
  • Develop simple chatbot.

Skills

Languages and Databases

Python
HTML5
CSS3
MySQL
PostgreSQL
Shell Scripting

Libraries

NumPy
Pandas
OpenCV
scikit-learn
matplotlib
NLTK

Frameworks

Django
Flask
Bootstrap
Keras
TensorFlow
PyTorch

Other

Git
AWS
Docker

Education

University of Wyoming

Laramie, WY

Degree: PhD in Computer Science
Area of Study: Causal Reasoning for Improving Generalization in Visual Question Answering

    Relevant Courseworks:

    • Intro to Artificial Intelligen
    • Machine Learning
    • High Perform Comput & Paradigm
    • Advanced Image Processing
    • Neural and Fuzzy Systems

Tarbiat Modares University

Tehran, Iran

Degree: Masters of Information Technology
CGPA: 3.68/4.0

    Relevant Courseworks:

    • Artificial Neural Networks
    • Neural and Fuzzy Systems

Contact