Open Source TypeScript Large Language Models (LLM)

TypeScript Large Language Models (LLM)

View 264 business solutions

Browse free open source TypeScript Large Language Models (LLM) and projects below. Use the toggles on the left to filter open source TypeScript Large Language Models (LLM) by OS, license, language, programming language, and project status.

  • Auth for GenAI | Auth0 Icon
    Auth for GenAI | Auth0

    Enable AI agents to securely access tools, workflows, and data with fine-grained control and just a few lines of code.

    Easily implement secure login experiences for AI Agents - from interactive chatbots to background workers with Auth0. Auth for GenAI is now available in Developer Preview
    Try free now
  • Gen AI apps are built with MongoDB Atlas Icon
    Gen AI apps are built with MongoDB Atlas

    Build gen AI apps with an all-in-one modern database: MongoDB Atlas

    MongoDB Atlas provides built-in vector search and a flexible document model so developers can build, scale, and run gen AI apps without stitching together multiple databases. From LLM integration to semantic search, Atlas simplifies your AI architecture—and it’s free to get started.
    Start Free
  • 1
    Flowise

    Flowise

    Drag & drop UI to build your customized LLM flow

    Open source UI visual tool to build your customized LLM flow using LangchainJS, written in Node Typescript/Javascript. Conversational agent for a chat model which utilizes chat-specific prompts and buffer memory. Open source is the core of Flowise, and it will always be free for commercial and personal usage. Flowise support different environment variables to configure your instance. You can specify the following variables in the .env file inside the packages/server folder.
    Downloads: 34 This Week
    Last Update:
    See Project
  • 2
    Langflow

    Langflow

    Low-code app builder for RAG and multi-agent AI applications

    Langflow is a low-code app builder for RAG and multi-agent AI applications. It’s Python-based and agnostic to any model, API, or database.
    Downloads: 33 This Week
    Last Update:
    See Project
  • 3
    Dify

    Dify

    One API for plugins and datasets, one interface for prompt engineering

    Dify is an easy-to-use LLMOps platform designed to empower more people to create sustainable, AI-native applications. With visual orchestration for various application types, Dify offers out-of-the-box, ready-to-use applications that can also serve as Backend-as-a-Service APIs. Unify your development process with one API for plugins and datasets integration, and streamline your operations using a single interface for prompt engineering, visual analytics, and continuous improvement. Out-of-the-box web sites supporting form mode and chat conversation mode A single API encompassing plugin capabilities, context enhancement, and more, saving you backend coding effort Visual data analysis, log review, and annotation for applications
    Downloads: 19 This Week
    Last Update:
    See Project
  • 4
    FreedomGPT

    FreedomGPT

    React and Electron-based app that executes the FreedomGPT LLM locally

    FreedomGPT is a locally executed large language model (LLM) application built using React and Electron, allowing users to interact with AI models privately on their Mac or Windows devices. The app enables offline operation, ensuring privacy and security while providing a chat-based interface for seamless communication with the AI. It supports integration with models like Liberty Edge and offers an open-source solution for those seeking more control over their AI interactions. The app's setup is simple, and it includes clear installation guides for both macOS and Windows platforms, as well as detailed instructions for building necessary libraries like llama.cpp.
    Downloads: 14 This Week
    Last Update:
    See Project
  • Build Securely on Azure with Proven Frameworks Icon
    Build Securely on Azure with Proven Frameworks

    Lay a foundation for success with Tested Reference Architectures developed by Fortinet’s experts. Learn more in this white paper.

    Moving to the cloud brings new challenges. How can you manage a larger attack surface while ensuring great network performance? Turn to Fortinet’s Tested Reference Architectures, blueprints for designing and securing cloud environments built by cybersecurity experts. Learn more and explore use cases in this white paper.
    Download Now
  • 5
    RWKV Runner

    RWKV Runner

    A RWKV management and startup tool, full automation, only 8MB

    RWKV (pronounced as RwaKuv) is an RNN with GPT-level LLM performance, which can also be directly trained like a GPT transformer (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, fast training, saves VRAM, "infinite" ctxlen, and free text embedding. Moreover it's 100% attention-free. Default configs has enabled custom CUDA kernel acceleration, which is much faster and consumes much less VRAM. If you encounter possible compatibility issues, go to the Configs page and turn off Use Custom CUDA kernel to Accelerate.
    Downloads: 9 This Week
    Last Update:
    See Project
  • 6
    Obsidian Text Generator Plugin

    Obsidian Text Generator Plugin

    Text generator is a handy plugin for Obsidian

    Text Generator is an open-source AI Assistant Tool that brings the power of Generative Artificial Intelligence to the power of knowledge creation and organization in Obsidian. For example, use Text Generator to generate ideas, attractive titles, summaries, outlines, and whole paragraphs based on your knowledge database.
    Downloads: 6 This Week
    Last Update:
    See Project
  • 7
    Code Review GPT

    Code Review GPT

    Your personal code reviewer powered by LLMs

    Code Review GPT uses Large Language Models to review code in your CI/CD pipeline. It helps streamline the code review process by providing feedback on code that may have issues or areas for improvement. Code Review GPT is in alpha and should be used for fun only. It may provide useful feedback but please check any suggestions thoroughly.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 8
    Lunary

    Lunary

    The production toolkit for LLMs. Observability, prompt management

    Lunary helps developers of LLM Chatbots develop and improve them.
    Downloads: 3 This Week
    Last Update:
    See Project
  • 9
    LangChain.js

    LangChain.js

    Building applications with LLMs through composability

    Building applications with LLMs through composability. Large language models (LLMs) are emerging as a transformative technology, enabling developers to build applications that they previously could not. But using these LLMs in isolation is often not enough to create a truly powerful app - the real power comes when you can combine them with other sources of computation or knowledge. This library is aimed at assisting in the development of those types of applications. This is built to integrate as seamlessly as possible with the LangChain Python package. Specifically, this means all objects (prompts, LLMs, chains, etc) are designed in a way where they can be serialized and shared between languages.
    Downloads: 2 This Week
    Last Update:
    See Project
  • Build Securely on AWS with Proven Frameworks Icon
    Build Securely on AWS with Proven Frameworks

    Lay a foundation for success with Tested Reference Architectures developed by Fortinet’s experts. Learn more in this white paper.

    Moving to the cloud brings new challenges. How can you manage a larger attack surface while ensuring great network performance? Turn to Fortinet’s Tested Reference Architectures, blueprints for designing and securing cloud environments built by cybersecurity experts. Learn more and explore use cases in this white paper.
    Download Now
  • 10
    WebLLM

    WebLLM

    Bringing large-language models and chat to web browsers

    WebLLM is a modular, customizable javascript package that directly brings language model chats directly onto web browsers with hardware acceleration. Everything runs inside the browser with no server support and is accelerated with WebGPU. We can bring a lot of fun opportunities to build AI assistants for everyone and enable privacy while enjoying GPU acceleration. WebLLM offers a minimalist and modular interface to access the chatbot in the browser. The WebLLM package itself does not come with UI, and is designed in a modular way to hook to any of the UI components. The following code snippet demonstrates a simple example that generates a streaming response on a webpage.
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    ChatGPT Admin Web

    ChatGPT Admin Web

    ChatGPT WebUI

    ChatGPT WebUI with user management and background management system. Deploy your commercial ChatGPT web application for free.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    Empirical

    Empirical

    Test and evaluate LLMs and model configurations

    Empirical is the fastest way to test different LLMs and model configurations, across all the scenarios that matter for your application.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    Langfuse

    Langfuse

    Open source LLM engineering platform: LLM Observability, metrics, etc.

    Langfuse is a logging and analytics tool for large language model (LLM) applications, providing insights into usage, performance, and troubleshooting.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 14
    Langtrace

    Langtrace

    Open Telemetry based end-to-end observability tool for LLM apps

    Langtrace is an open-source, Open Telemetry based end-to-end observability tool for LLM applications, providing real-time tracing, evaluations, and metrics for popular LLMs, LLM frameworks, vectors, and more.. Integrate using Typescript, and Python. Langtrace is an open-source observability tool that collects and analyzes traces and metrics to help you improve your LLM apps.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 15
    Pezzo

    Pezzo

    Open-source, developer-first LLMOps platform

    Pezzo enables you to build, test, monitor and instantly ship AI all in one platform, while constantly optimizing for cost and performance. Packed with powerful features to streamline your workflow, so you can focus on what matters. Pezzo is a fully cloud-native and open-source LLMOps platform. Seamlessly observe and monitor your AI operations, troubleshoot issues, save up to 90% on costs and latency, collaborate and manage your prompts in one place, and instantly deliver AI changes.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 16
    Quivr

    Quivr

    Your Second Brain supercharged by Generative AI

    Quivr, your second brain, utilizes the power of GenerativeAI to store and retrieve unstructured information. Think of it as Obsidian, but turbocharged with AI capabilities.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 17
    SkyAGI

    SkyAGI

    SkyAGI: Emerging human-behavior simulation capability in LLM

    SkyAGI is a python package that demonstrates LLM's emerging capability in simulating believable human behaviors. Specifically, SkyAGI implements the idea of Generative Agents and delivers a role-playing game that creates a very interesting user experience. Different from previous AI-based NPC systems, SkyAGI's NPC generates very believable human responses. The interesting observations in this demo show a huge potential for rethinking game development in many aspects, such as NPC script writing.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 18
    Vercel AI SDK

    Vercel AI SDK

    Build AI-powered applications with React, Svelte, Vue, and Solid

    The Vercel AI SDK is a library for building AI-powered streaming text and chat UIs.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 19
    promptfoo

    promptfoo

    Evaluate and compare LLM outputs, catch regressions, improve prompts

    Ensure high-quality LLM outputs with automatic evals. Use a representative sample of user inputs to reduce subjectivity when tuning prompts. Use built-in metrics, LLM-graded evals, or define your own custom metrics. Compare prompts and model outputs side-by-side, or integrate the library into your existing test/CI workflow. Use OpenAI, Anthropic, and open-source models like Llama and Vicuna, or integrate custom API providers for any LLM API.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 20
    react-llm

    react-llm

    Easy-to-use headless React Hooks to run LLMs in the browser with WebGP

    Easy-to-use headless React Hooks to run LLMs in the browser with WebGPU. As simple as useLLM().
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
Want the latest updates on software, tech news, and AI?
Get latest updates about software, tech news, and AI from SourceForge directly in your inbox once a month.