TripoSR icon

TripoSR

Open Source
4.5
Stability AI & Tripo

TripoSR is a fast feed-forward 3D reconstruction model jointly developed by Stability AI and Tripo AI that generates detailed 3D meshes from single input images in under one second. Unlike optimization-based methods that require minutes of processing per object, TripoSR uses a transformer-based architecture built on the Large Reconstruction Model framework to predict 3D geometry directly from a single 2D photograph in a single forward pass. The model accepts any standard image as input and produces a textured 3D mesh suitable for use in game engines, 3D modeling software, and augmented reality applications. TripoSR excels at reconstructing everyday objects, furniture, vehicles, characters, and organic shapes with impressive geometric accuracy and surface detail. Released under the MIT license in March 2024, the model is fully open source and can run on consumer-grade GPUs without specialized hardware. It supports batch processing for efficient conversion of multiple images and integrates seamlessly with popular 3D pipelines including Blender, Unity, and Unreal Engine. The model is particularly valuable for game developers, product designers, and e-commerce teams who need rapid 3D asset creation from product photographs. Output meshes can be exported in OBJ and GLB formats with configurable resolution settings. TripoSR represents a significant step toward democratizing 3D content creation by making high-quality reconstruction accessible without expensive scanning equipment or manual modeling expertise.

Text to 3D
Image to 3D

Key Highlights

Sub-Second 3D Generation Speed

Generates complete textured 3D meshes from a single image in under 0.5 seconds on modern GPUs through its feed-forward architecture without iterative optimization

Production-Ready Mesh Output

Produces 3D meshes with texture maps in standard formats like OBJ and GLB, providing immediately usable assets for games, AR/VR, and 3D applications

LRM-Based Transformer Architecture

Built on the Large Reconstruction Model framework using triplane neural radiance fields, achieving high-quality reconstruction through a single efficient forward pass

MIT License Commercial Freedom

Released under the permissive MIT license by Stability AI and Tripo AI, allowing unrestricted commercial deployment and integration without licensing fees

About

TripoSR is a fast feed-forward 3D reconstruction model jointly developed by Stability AI and Tripo AI that generates detailed 3D meshes from single input images in under one second. Released in March 2024, TripoSR represents a significant advancement in single-image 3D reconstruction by eliminating the need for time-consuming per-shape optimization that characterizes many competing approaches. The model bridges the gap between academic research and production environments by offering a practical solution for industrial-scale 3D asset generation.

The model architecture is based on the Large Reconstruction Model (LRM) framework, using a transformer-based design that processes the input image through a vision encoder and generates a triplane-based neural radiance field representation. This triplane representation is then converted to a textured 3D mesh through marching cubes extraction. The entire pipeline runs in a single forward pass without iterative optimization, enabling generation speeds of under 0.5 seconds on modern GPUs. The DINOv2 vision encoder extracts rich semantic and structural features from the input image, enhancing reconstruction quality and enabling consistent performance across different object categories.

TripoSR produces 3D meshes with corresponding texture maps, providing immediately usable assets for 3D applications. The output quality captures the geometry and appearance of the input subject with reasonable accuracy, including details like surface textures, color variations, and overall shape proportions. The model successfully handles a variety of input types including product photos, character images, artwork, and object photographs. It achieves particularly high-fidelity results on objects with smooth surfaces and distinct silhouettes, while complex scenes with fine-grained details or transparent objects may exhibit quality limitations.

The feed-forward architecture means TripoSR scales efficiently for batch processing scenarios, as each reconstruction takes a fixed amount of time regardless of object complexity. This characteristic makes it particularly suitable for applications requiring rapid 3D asset generation at scale, such as e-commerce product catalogs, game development prototyping, and AR/VR content pipelines. The capacity to process thousands of objects per hour on a single consumer GPU significantly reduces costs in industrial use cases and decreases the need for manual 3D modeling compared to traditional approaches.

In terms of training data, TripoSR was trained on the Objaverse dataset, and the diversity of objects in this dataset directly influences the model's generalization capability. The model demonstrates strong performance on objects within its training distribution while potentially showing degraded results on inputs with unusual geometries or rare object categories. Output meshes can be exported in OBJ and GLB formats and are fully compatible with standard 3D software including Blender, Unity, and Unreal Engine. Mesh resolution and texture size are user-configurable to accommodate different application requirements.

Released under the MIT license, TripoSR is fully open-source and available for both research and commercial use. The model is accessible through Hugging Face with pre-trained weights and can be run locally on consumer GPUs. Its combination of speed, quality, and open licensing has made it one of the most popular open-source single-image 3D reconstruction tools available. Community-developed integrations and extensions have expanded the model's reach to ComfyUI plugins, Gradio-based web applications, and automated 3D asset production pipelines.

Use Cases

1

E-Commerce 3D Product Catalogs

Rapidly convert product photography into 3D models for interactive product viewers, AR try-on experiences, and 3D e-commerce listings

2

Game Development Asset Prototyping

Generate quick 3D mesh prototypes from concept art and reference images for game development blocking and level design iteration

3

AR/VR Content Pipeline

Feed images into automated pipelines to generate 3D assets for augmented reality and virtual reality applications at scale

4

3D Printing Model Generation

Create printable 3D meshes from photographs of objects for rapid prototyping, collectibles, and custom manufacturing applications

Pros & Cons

Pros

  • Generates 3D models in under 0.5 seconds on NVIDIA A100 GPU — exceptionally fast single-image reconstruction
  • Outperforms other open-source alternatives in both qualitative and quantitative evaluations across multiple datasets
  • Released under MIT license with source code, pretrained models, and interactive online demo
  • Minimal learning curve — requires only 1-2 hours to get started
  • Produces clean, usable mesh output suitable for downstream 3D applications

Cons

  • Single-view ambiguity causes inaccuracies when inferring hidden geometry, especially for complex shapes
  • Fine surface details, textures, and intricate patterns are often missing or smoothed over
  • Highly dependent on input image quality — poorly lit or ambiguous images produce subpar results
  • Struggles with highly intricate objects or scenes with significant occlusion
  • Requires clean background or transparent PNG for best results — real-world photos need preprocessing

Technical Details

Parameters

N/A

License

MIT

Features

  • Single Image to 3D Mesh
  • Sub-Second Generation Speed
  • Feed-Forward Architecture
  • No Per-Shape Optimization
  • Multiple Output Formats (OBJ, GLB)
  • Texture Map Generation
  • MIT Open-Source License
  • Hugging Face Integration

Benchmark Results

MetricValueCompared ToSource
Üretim Süresi~0.5 saniye (A100)Shap-E: ~10sTripoSR Paper / Stability AI Blog
F-Score (GSO Dataset)0.477 (F-Score@0.1)LGM: 0.413TripoSR Paper (arXiv:2403.02151)
Mesh Kalitesi (Vertex Sayısı)~50K-200K vertices (marching cubes)Shap-E: ~4K verticesTripoSR GitHub / Hugging Face
Texture Çözünürlüğü1024x1024Shap-E: vertex colors onlyTripoSR GitHub

Available Platforms

hugging face
replicate
fal ai

News & References

Frequently Asked Questions

Related Models

TRELLIS icon

TRELLIS

Microsoft Research|Unknown

TRELLIS is a revolutionary AI model developed by Microsoft Research that generates high-quality 3D assets from text descriptions or single 2D images using a novel Structured Latent Diffusion architecture. Released in December 2024, TRELLIS represents a fundamental advancement in 3D content generation by operating in a structured latent space that encodes geometry, texture, and material properties simultaneously rather than treating them as separate stages. The model produces complete 3D meshes with detailed PBR (Physically Based Rendering) textures, enabling direct use in game engines, 3D rendering pipelines, and AR/VR applications without extensive manual post-processing. TRELLIS supports both text-to-3D generation where users describe desired objects in natural language and image-to-3D reconstruction where a single photograph is converted into a full 3D model with inferred geometry from occluded viewpoints. The structured latent representation ensures geometric consistency and prevents the common artifacts seen in other 3D generation approaches such as floating geometry, texture seams, and unrealistic proportions. TRELLIS outputs standard 3D formats including GLB and OBJ with UV-mapped textures, making integration with professional tools like Blender, Unity, and Unreal Engine straightforward. Released under the MIT license, the model is fully open source and available on GitHub. Key applications include rapid 3D asset prototyping for game development, architectural visualization, product design mockups, virtual staging for real estate, educational 3D content creation, and metaverse asset generation. The model particularly benefits indie developers and small studios who lack resources for traditional 3D modeling workflows.

Open Source
4.5
Meshy icon

Meshy

Meshy AI|N/A

Meshy is a proprietary AI-powered 3D generation platform developed by Meshy AI that creates detailed, production-ready 3D models from text descriptions and images. The platform combines text-to-3D and image-to-3D capabilities with advanced AI texturing features, positioning itself as a comprehensive solution for rapid 3D content creation. Meshy uses a transformer-based architecture that generates textured 3D meshes with PBR-compatible materials, making outputs directly usable in game engines like Unity and Unreal Engine without additional processing. The platform offers multiple generation modes including text-to-3D for creating objects from written descriptions, image-to-3D for converting photographs into 3D models, and AI texturing for applying realistic materials to existing untextured meshes. Generated models include proper UV mapping, normal maps, and physically based rendering materials suitable for professional workflows. Meshy provides both a web-based interface and an API for programmatic access, making it accessible to individual artists and scalable for enterprise pipelines. The platform is particularly popular among game developers, animation studios, and AR/VR content creators who need to produce large volumes of 3D assets efficiently. As a proprietary commercial service launched in 2023, Meshy operates on a subscription model with free tier access for limited generations. The platform continuously updates its models to improve output quality, topology optimization, and texture fidelity, competing directly with other AI 3D generation services in the rapidly evolving market.

Proprietary
4.4
InstantMesh icon

InstantMesh

Tencent|N/A

InstantMesh is a feed-forward 3D mesh generation model developed by Tencent that creates high-quality textured 3D meshes from single input images through a multi-view generation and sparse-view reconstruction pipeline. Released in April 2024 under the Apache 2.0 license, InstantMesh combines a multi-view diffusion model with a large reconstruction model to achieve both speed and quality in single-image 3D reconstruction. The pipeline first generates multiple consistent views of the input object using a fine-tuned multi-view diffusion model, then feeds these views into a transformer-based reconstruction network that predicts a triplane neural representation, which is finally converted to a textured mesh. This two-stage approach produces significantly higher quality results than single-stage methods while maintaining generation times of just a few seconds. InstantMesh supports both text-to-3D workflows when combined with an image generation model and direct image-to-3D conversion from photographs or artwork. The output meshes include detailed geometry and texture maps compatible with standard 3D software and game engines. The model handles a wide variety of object types including characters, vehicles, furniture, and organic shapes with good geometric fidelity. As an open-source project with code and weights available on GitHub and Hugging Face, InstantMesh has become a popular choice for developers building 3D asset generation pipelines. It is particularly useful for game development, e-commerce product visualization, and rapid prototyping scenarios where fast turnaround and reasonable quality are both important requirements.

Open Source
4.3
Shap-E icon

Shap-E

OpenAI|N/A

Shap-E is a 3D generation model developed by OpenAI that creates 3D objects directly from text descriptions or input images by generating the parameters of implicit neural representations. Unlike its predecessor Point-E which produces point clouds, Shap-E generates Neural Radiance Fields (NeRF) and textured meshes that can be directly rendered and used in 3D applications. The model employs a two-stage training approach where an encoder first learns to map 3D assets to implicit function parameters, then a conditional diffusion model learns to generate those parameters from text or image inputs. This architecture enables fast generation times of just a few seconds on a modern GPU. Shap-E supports both text-to-3D and image-to-3D workflows, making it versatile for different creative pipelines. The generated 3D objects include color and texture information, producing more complete results than geometry-only approaches. Released under the MIT license in May 2023, the model is fully open source with pre-trained weights available on GitHub. While the output quality may not match optimization-heavy methods like DreamFusion that take minutes per object, Shap-E offers a practical balance between speed and quality for rapid prototyping and concept exploration. The model is particularly useful for game developers, 3D artists, and researchers who need quick 3D visualizations from text prompts. As one of OpenAI's contributions to open-source 3D AI research, Shap-E has influenced subsequent work in fast feed-forward 3D generation approaches.

Open Source
4.0

Quick Info

ParametersN/A
Typetransformer
LicenseMIT
Released2024-03
Rating4.5 / 5
CreatorStability AI & Tripo

Links

Tags

triposr
3d
fast
reconstruction
Visit Website