Skip to content
View Masoudjafaripour's full-sized avatar
🎯
Focusing
🎯
Focusing

Highlights

  • Pro

Organizations

@Adaptive-Robotic-Lab

Block or report Masoudjafaripour

Block user

Prevent this user from interacting with your repositories and sending you notifications. Learn more about blocking users.

You must be logged in to block users.

Maximum 250 characters. Please don't include any personal information such as legal names or email addresses. Markdown supported. This note will be visible to only you.
Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse
Masoudjafaripour/README.md

I am a grad student in CS at the UofA, supervised by Prof. Osmar Zaiane. My research focuses on multimodal generative models, including LLMs, VLMs, and diffusion models, with emphasis on 3D spatial reasoning. Current work includes scaling 3D spatial reasoning in multimodal image generation with test-time scalling, multimodal CoT, applying post-training methods (SFTv& RL Fine-tuning), and curating vision-language datasets.


📊 Current Focus

  • Spatial & Visual Reasoning with LLMs/VLMs/MMLMs
  • Vision-Language Understanding & Embodied Spatial Reasoning
  • 3D Representations, Grounding, & Space Understanding
  • Building Vision-Language Datasets for Embodied Multi-Agent Systems
  • Visual and Geometry Retrieval Systems

🎓 Academic Background

  • M.Sc. in CS, University of Alberta (Present)
  • Ph.D. in ECE, University of Alberta (Transferred to CS)
  • M.Sc. & B.Sc. in ME, Sharif University of Technology & Univ. of Tehran

💬 Connect with Me

Website Badge Twitter Badge LinkedIn Badge ResearchGate Badge Google Scholar Badge

Pinned Loading

  1. AIFP AIFP Public

    Adaptive Iterative Feedback Prompting for Obstacle-Aware Path Planning via LLMs - LM4Planning - AAAI2025

    Jupyter Notebook 6 3

  2. FM_RL_Survey FM_RL_Survey Public

    A repo for survey paper "The Evolving Landscape of LLM- and VLM-Integrated Reinforcement Learning" and a collection of AWESOME papers focused on using LLMs, VLMs for improving RL.

    5

  3. nanochat-VLM nanochat-VLM Public

    A minimal, hackable Vision-Language Model built on Karpathy’s nanochat — add image understanding and multimodal chat for under $200 in compute.

    Python 6

  4. Single_Agent_Search_Project Single_Agent_Search_Project Public template

    Forked from gppc-dev/startkit-classic

    This repository is for Single Agent Course Project (3rd Assignment): Evaluating FastMap and Hybrid Heuristics for Fast Grid Pathfinding

    C++ 4 1

  5. Awesome-LLM-Planning-Spatial-Reasoning Awesome-LLM-Planning-Spatial-Reasoning Public

    Awesome-LLM-Planning

    5

  6. OnlineRLHF OnlineRLHF Public

    A repo for Implemented online preference-based reward learning under human irrationality & delayed feedback

    Python 5 1