Skip to content
  • Product
    • MoAI Inference Framework
    • Moreh vLLM
    • MoAI Training Framework
    • MoAI Platform
  • Solutions

    Infrastructure

    AMD GPU Appliance

    Deploy fully integrated AMD GPU-based cluster systems with scalable RoCE networking

    Tenstorrent Appliance

    Deliver the lowest TCO with inherently scalable, network-integrated chips

    Use Cases

    End-to-End Model Deployment

    Build cost-effective inference endpoints on-premises or in the cloud

    Large Scale Training

    Maximize GPU utilization and reduce training costs at the 1,000+ GPU scale

    Operation

    GPU Virtualization

    Flexible GPU aggregation, decomposition, and scaling with heterogeneous GPU support

    System Reliability

    Automatic GPU failover and diagnostic hardware monitoring

    • AMD GPU Appliance
    • Tenstorrent Appliance
    • End-to-End Model Deployment
    • Large Scale Training
    • GPU Virtualization
    • System Reliability
  • Resources
    • Blog
    • Docs
    • Demo Videos
    • Open Source
  • Career
  • Company
    • About
    • Contact
    • Newsroom

morehiodev

  • Moreh-Tenstorrent AI Data Center Solution System Architecture

    November 18, 2025

    Moreh combine Tenstorrent’s lightweight and scalable hardware with our proprietary software stack to deliver an efficient and flexible solution for large-scale AI data centers.

  • Moreh and Tenstorrent Unveil Scalable, Cost-Efficient AI Data Center Solution at SuperComputing 2025

    November 17, 2025

    Moreh, a provider of optimized AI infrastructure software, and Tenstorrent, an AI semiconductor company, are unveiling a scalable, cost-efficient AI data center solution at SuperComputing 2025 in St. Louis, Missouri.

  • 21K Output Tokens Per Second DeepSeek Inference on AMD Instinct MI300X GPUs with Expert Parallelism

    November 13, 2025

    Moreh demonstrated that DeepSeek-R1 inference can be executed at a decoding throughput of >21,000 tokens/sec by implementing EP on the ROCm software stack.

  • Runtime Draft Model Training: Adapting Speculative Decoding to Real-World Workloads

    November 10, 2025

    TIDE provides a method to optimize inference computation on newer GPUs by utilizing older or idle GPUs for runtime draft model training, resulting in better overall cost-performance at the system level.

  • Cost-Efficient AI at Scale is a Software Problem

    November 7, 2025

    EE Times — Cost-efficient AI at scale is a software problem, given that all Nvidia competitors are lagging behind on software development. Not only that, but software will also become even more critical as heterogeneity in the data center becomes commonplace.

  • Distributed Inference on Heterogeneous Accelerators Including GPUs, Rubin CPX, and AI Accelerators

    September 23, 2025

    MoAI Inference Framework supports automatic and efficient distributed inference on heterogeneous accelerators such as AMD MI300X + MI308X and NVIDIA Rubin CPX + GPU.

  • Moreh and SGLang team up to showcase distributed inference system on AMD at AI Infra Summit 2025

    September 11, 2025

    Introducing distributed inference systems on AMD with greater efficiency than NVIDIA, and unveiling collaborations with Tenstorrent and SGLang.

  • Moreh vLLM Performance Evaluation: Llama 3.3 70B on AMD Instinct MI300X GPUs

    August 30, 2025

    Moreh vLLM achieves 1.68x higher output TPS, 2.02x lower TTFT, and 1.59x lower TPOT compared to the original vLLM for Meta's Llama 3.3 70B model.

  • Moreh vLLM Performance Evaluation: DeepSeek V3/R1 671B on AMD Instinct MI300X GPUs

    August 29, 2025

    Moreh vLLM achieves 1.68x higher output TPS, 1.75x lower TTFT, and 1.70x lower TPOT compared to the original vLLM for the DeepSeek V3/R1 671B model.

  • DeepSeek V3 and R1 on MoAI: 1. Fine-Tuning on AMD GPU Clusters

    February 20, 2025

    MoAI provides a PyTorch-compatible environment that makes LLM fine-tuning on hundreds of AMD GPUs super easy, including DeepSeek 671B MoE.

12Next

Moreh, Inc.

  • Home
  • About
  • Career
  • Contact
  • Docs
  • Blog
  • Newsroom
  • Privacy Policy
  • Terms of Use
  • Home
  • About
  • Career
  • Contact
  • Docs
  • Blog
  • Newsroom

© 2025 Moreh, Inc. All right reserved.

Page load link
Go to Top