Kenneth Xu

About Me

Hello! I'm Kenneth Xu, a senior graduating in December 2025 with a double major in Computer Science and Cognitive Science at the University of Michigan. I'm passionate in bridging technology and human perception—whether I’m building full-stack web apps or exploring AI’s creative side.


I'm focused on conducting research on 3D Segmentation and semantic representations of 3D Gaussian Splats, working with peers at USC and mentored by Professor Chen Jiasi. I'm also currently a member of Professor Xu Wang's Lifelong Learning Lab where I'm working on enhancing Human Computer Interaction.


When I'm not in class or coding, you can find me playing competitive table tennis, cooking up favorites like omurice and steak, lifting, or playing basketball.

Kenneth Xu

Research Projects

MastSAM Preview
Multi-view Segmentation

MastSAM: Solving Multi-view Inconsistency in 3D Segmentation

IJCNN Publication, 2025

We tackle the challenge of inconsistent segmentation outputs from 2D models by leveraging a 3D coordinate mapping technique. MastSAM effectively aligns segmentation masks across multiple views, ensuring coherent and consistent 3D segmentation results across subsequent image frames. [PDF]

Tiny-YOLOSAM Preview
Hybrid Instance Segmentation

Tiny-YOLOSAM: YOLO + TinySAM Hybrid Instance Segmentation

Research Project, 2025

YOLOv12 object detection with TinySAM (efficient Segment Anything Model) for fast instance segmentation. We implement and evaluate two novel approaches: YOLO-only (YOLOv12 boxes → TinySAM segmentation), and Hybrid (YOLO for foreground + sparse points for background). The Hybrid approach achieves 3-8× faster inference than existingmethods while maintaining competitive AR/mIoU metrics, balancing speed and coverage. [GitHub Repository]

Research Interests

My ultimate vision is a world enabled by AI where anyone can travel and perceive anywhere in the world without being physically present—imagine visiting the Roman Colosseum and witnessing gladiatorial combat unfold before your eyes, or exploring the depths of the Amazon rainforest with perfect fidelity. This dream drives my research in 3D scene understanding and semantic representation, where I develop foundational technologies that make such immersive experiences possible.


Through my work on MastSAM and Splat Feature Solver, I'm building the technical foundations for this vision. MastSAM solves the critical challenge of multi-view inconsistency in 3D segmentation by leveraging 3D coordinate mapping techniques, ensuring coherent spatial understanding across different perspectives—essential for creating seamless virtual environments. Meanwhile, Splat Feature Solver enables rich semantic understanding of 3D scenes by efficiently lifting powerful image descriptors like CLIP into 3D Gaussian Splat representations, allowing AI systems to understand and interact with virtual worlds at a semantic level.


These technologies converge toward creating photorealistic, semantically-aware virtual worlds where users can not only see but truly understand and interact with their surroundings. My research in Human-Computer Interaction complements this vision by designing intuitive interfaces that make these complex 3D environments accessible to everyone, bridging the gap between cutting-edge AI capabilities and natural human interaction.


I'm always eager to connect with fellow researchers, industry professionals, or anyone passionate about making virtual reality indistinguishable from physical reality. Whether you're interested in collaborating on 3D scene understanding, semantic representation, or building the interfaces that will define our virtual future, I'd love to hear from you!

Personal Projects

LocationAlert: Geofence-Based Location Monitoring

Personal Project, 2024

A mobile application that alerts users when they leave a predefined geofenced location. Designed for ensuring location-based safety and reminders, this project integrates GPS tracking and push notifications to keep users informed about their location status in real time. [GitHub Repository]

Education

University of Michigan, Ann Arbor

Bachelor of Science in Computer Science & Cognitive Science

Expected Graduation: December 2025

Relevant Coursework: Machine Learning, Computer Vision, Foundations of Computer Science,Programming & Data Structures, Computer Organization, AI-Enabled XR, Autonomous Robotics, Practical Data Science, Discrete Mathematics, Statistics

Research Focus: 3D Scene Understanding, Semantic Representation, Gaussian Splatting, Human-Computer Interaction in Education

Technical Skills

Programming Languages

Python, C++, JavaScript, TypeScript, Java, MATLAB, R

ML/AI Frameworks

PyTorch, TensorFlow, OpenCV, CLIP, SAM, Gaussian Splatting, 3DGS

Research Tools

Git, Conda, Docker, Viser, Blender, Unity, Jupyter Notebooks

Web Development

React, Node.js, HTML/CSS, REST APIs, Database Design

Updates

  • September 2025: Submitted Splat Feature Solver to ICLR 2026
  • July 2025: Presented MastSAM at IJCNN 2025
  • January 2025: Joined Professor Xu's Life-long Learning Lab as a Research Assistant
  • August 2024: Wrapped up an internship at Nowadays (YC S23), contributing to frontend development and database automation.
  • March 2024: Completed an internship at Hemingway, developing AI-driven healthcare tools, getting experience in NLP and HCI.