Results for: pytorch

Suggested Categories:

Artificial Intelligence Software
Artificial Intelligence (AI) software is computer technology designed to simulate human intelligence. It can be used to perform tasks that require cognitive abilities, such as problem-solving, data analysis, visual perception and language translation. AI applications range from voice recognition and virtual assistants to autonomous vehicles and medical diagnostics.

42 Products for "pytorch"

  • 1
    PyTorch

    PyTorch

    PyTorch

    ...Scalable distributed training and performance optimization in research and production is enabled by the torch-distributed backend. A rich ecosystem of tools and libraries extends PyTorch and supports development in computer vision, NLP and more. PyTorch is well supported on major cloud platforms, providing frictionless development and easy scaling. Select your preferences and run the install command. Stable represents the most currently tested and supported version of PyTorch. This should be suitable for many users. ...
  • 2
    TorchMetrics

    TorchMetrics

    TorchMetrics

    TorchMetrics is a collection of 90+ PyTorch metrics implementations and an easy-to-use API to create custom metrics. A standardized interface to increase reproducibility. It reduces boilerplate. distributed-training compatible. It has been rigorously tested. Automatic accumulation over batches. Automatic synchronization between multiple devices. You can use TorchMetrics in any PyTorch model, or within PyTorch Lightning to enjoy additional benefits.
    Starting Price: Free
  • 3
    AWS Neuron

    AWS Neuron

    Amazon Web Services

    ...For distributed model training, the Neuron SDK supports libraries, such as Megatron-LM and PyTorch Fully Sharded Data Parallel (FSDP).
  • 4
    DeepSpeed

    DeepSpeed

    Microsoft

    ...It can also train up to 13 billion parameters in a single GPU. DeepSpeed is developed by Microsoft and aims to offer distributed training for large-scale models. It's built on top of PyTorch, which specializes in data parallelism.
    Starting Price: Free
  • 5
    Google Cloud Deep Learning VM Image
    ...Deep Learning VM Image makes it easy and fast to instantiate a VM image containing the most popular AI frameworks on a Google Compute Engine instance without worrying about software compatibility. You can launch Compute Engine instances pre-installed with TensorFlow, PyTorch, scikit-learn, and more. You can also easily add Cloud GPU and Cloud TPU support. Deep Learning VM Image supports the most popular and latest machine learning frameworks, like TensorFlow and PyTorch. To accelerate your model training and deployment, Deep Learning VM Images are optimized with the latest NVIDIA® CUDA-X AI libraries and drivers and the Intel® Math Kernel Library. ...
  • 6
    Bayesforge

    Bayesforge

    Quantum Programming Studio

    Bayesforge™ is a Linux machine image that curates the very best open source software for the data scientist who needs advanced analytical tools, as well as for quantum computing and computational mathematics practitioners who seek to work with one of the major QC frameworks. The image combines common machine learning frameworks, such as PyTorch and TensorFlow, with open source software from D-Wave, Rigetti as well as the IBM Quantum Experience and Google's new quantum computing language Cirq, as well as other advanced QC frameworks. For instance our quantum fog modeling framework, and our quantum compiler Qubiter which can cross-compile to all major architectures. All software is made accessible through the Jupyter WebUI which, due to its modular architecture, allows the user to code in Python, R, and Octave.
  • 7
    Azure Databricks
    Unlock insights from all your data and build artificial intelligence (AI) solutions with Azure Databricks, set up your Apache Spark™ environment in minutes, autoscale, and collaborate on shared projects in an interactive workspace. Azure Databricks supports Python, Scala, R, Java, and SQL, as well as data science frameworks and libraries including TensorFlow, PyTorch, and scikit-learn. Azure Databricks provides the latest versions of Apache Spark and allows you to seamlessly integrate with open source libraries. Spin up clusters and build quickly in a fully managed Apache Spark environment with the global scale and availability of Azure. Clusters are set up, configured, and fine-tuned to ensure reliability and performance without the need for monitoring. ...
  • 8
    GPUonCLOUD

    GPUonCLOUD

    GPUonCLOUD

    ...However, with GPUonCLOUD’s dedicated GPU servers, it's a matter of hours. You may want to opt for pre-configured systems or pre-built instances with GPUs featuring deep learning frameworks like TensorFlow, PyTorch, MXNet, TensorRT, libraries e.g. real-time computer vision library OpenCV, thereby accelerating your AI/ML model-building experience. Among the wide variety of GPUs available to us, some of the GPU servers are best fit for graphics workstations and multi-player accelerated gaming. Instant jumpstart frameworks increase the speed and agility of the AI/ML environment with effective and efficient environment lifecycle management.
    Starting Price: $1 per hour
  • 9
    RoBERTa
    RoBERTa builds on BERT’s language masking strategy, wherein the system learns to predict intentionally hidden sections of text within otherwise unannotated language examples. RoBERTa, which was implemented in PyTorch, modifies key hyperparameters in BERT, including removing BERT’s next-sentence pretraining objective, and training with much larger mini-batches and learning rates. This allows RoBERTa to improve on the masked language modeling objective compared with BERT and leads to better downstream task performance. We also explore training RoBERTa on an order of magnitude more data than BERT, for a longer amount of time. ...
    Starting Price: Free
  • 10
    Daft

    Daft

    Daft

    ...Its familiar Python dataframe API is built to outperform Spark in performance and ease of use. Daft plugs directly into your ML/AI stack through efficient zero-copy integrations with essential Python libraries such as Pytorch and Ray. It also allows requesting GPUs as a resource for running models. Daft runs locally with a lightweight multithreaded backend. When your local machine is no longer sufficient, it scales seamlessly to run out-of-core on a distributed cluster. Daft can handle User-Defined Functions (UDFs) in columns, allowing you to apply complex expressions and operations to Python objects with the full flexibility required for ML/AI. ...
  • 11
    Flower

    Flower

    Flower

    ...It enables training on data located on devices or servers without transferring the data itself, thereby enhancing privacy and reducing bandwidth usage. Flower supports a wide range of machine learning frameworks, including PyTorch, TensorFlow, Hugging Face Transformers, scikit-learn, and XGBoost, and is compatible with various platforms and cloud services like AWS, GCP, and Azure. It offers flexibility through customizable strategies and supports both horizontal and vertical federated learning scenarios. Flower's architecture allows for scalable experiments, with the capability to handle workloads involving tens of millions of clients. ...
    Starting Price: Free
  • 12
    NVIDIA FLARE
    ...It enables secure, privacy-preserving AI model training by allowing multiple parties to collaboratively train models without sharing raw data. FLARE supports various machine learning frameworks such as PyTorch, TensorFlow, RAPIDS, and XGBoost, making it adaptable to existing workflows. FLARE's componentized architecture allows for customization and scalability, supporting both horizontal and vertical federated learning. It is suitable for applications requiring data privacy and regulatory compliance, such as medical imaging and financial analytics. ...
    Starting Price: Free
  • 13
    Fabric for Deep Learning (FfDL)
    Deep learning frameworks such as TensorFlow, PyTorch, Caffe, Torch, Theano, and MXNet have contributed to the popularity of deep learning by reducing the effort and skills needed to design, train, and use deep learning models. Fabric for Deep Learning (FfDL, pronounced “fiddle”) provides a consistent way to run these deep-learning frameworks as a service on Kubernetes. The FfDL platform uses a microservices architecture to reduce coupling between components, keep each component simple and as stateless as possible, isolate component failures, and allow each component to be developed, tested, deployed, scaled, and upgraded independently. ...
  • 14
    Cerebrium

    Cerebrium

    Cerebrium

    Deploy all major ML frameworks such as Pytorch, Onnx, XGBoost etc with 1 line of code. Don't have your own models? Deploy our prebuilt models that have been optimised to run with sub-second latency. Fine-tune smaller models on particular tasks in order to decrease costs and latency while increasing performance. It takes just a few lines of code and don't worry about infrastructure, we got it.
    Starting Price: $ 0.00055 per second
  • 15
    Amazon Elastic Inference
    Amazon Elastic Inference allows you to attach low-cost GPU-powered acceleration to Amazon EC2 and Sagemaker instances or Amazon ECS tasks, to reduce the cost of running deep learning inference by up to 75%. Amazon Elastic Inference supports TensorFlow, Apache MXNet, PyTorch and ONNX models. Inference is the process of making predictions using a trained model. In deep learning applications, inference accounts for up to 90% of total operational costs for two reasons. Firstly, standalone GPU instances are typically designed for model training - not for inference. While training jobs batch process hundreds of data samples in parallel, inference jobs usually process a single input in real time, and thus consume a small amount of GPU compute. ...
  • 16
    CodeT5

    CodeT5

    Salesforce

    Code for CodeT5, a new code-aware pre-trained encoder-decoder model. Identifier-aware unified pre-trained encoder-decoder models for code understanding and generation. This is the official PyTorch implementation for the EMNLP 2021 paper from Salesforce Research. CodeT5-large-ntp-py is specially optimized for Python code generation tasks and employed as the foundation model for our CodeRL, yielding new SOTA results on the APPS Python competition-level program synthesis benchmark. This repo provides the code for reproducing the experiments in CodeT5. ...
  • 17
    Horovod

    Horovod

    Horovod

    ...Horovod can additionally run on top of Apache Spark, making it possible to unify data processing and model training into a single pipeline. Once Horovod has been configured, the same infrastructure can be used to train models with any framework, making it easy to switch between TensorFlow, PyTorch, MXNet, and future frameworks as machine learning tech stacks continue to evolve.
    Starting Price: Free
  • 18
    Groq

    Groq

    Groq

    ...Additionally, eliminating external memory bottlenecks enables the LPU inference engine to deliver orders of magnitude better performance on LLMs compared to GPUs. Groq supports standard machine learning frameworks such as PyTorch, TensorFlow, and ONNX for inference.
  • 19
    Keepsake

    Keepsake

    Replicate

    ...This facilitates the retrieval of code and weights from any checkpoint, aiding in re-training or model deployment. Keepsake supports various machine learning frameworks, including TensorFlow, PyTorch, scikit-learn, and XGBoost, by saving files and dictionaries in a straightforward manner. It also offers features such as experiment comparison, enabling users to analyze differences in parameters, metrics, and dependencies across experiments.
    Starting Price: Free
  • 20
    Hugging Face Transformers
    ...Simple and optimized inference class for many machine learning tasks like text generation, image segmentation, automatic speech recognition, document question answering, and more. A comprehensive trainer that supports features such as mixed precision, torch.compile, and FlashAttention for training and distributed training for PyTorch models.​ Fast text generation with large language models and vision language models. Every model is implemented from only three main classes (configuration, model, and preprocessor) and can be quickly used for inference or training.
    Starting Price: $9 per month
  • 21
    SiMa

    SiMa

    SiMa

    ...The MLSoC integrates multiple modalities, including text, image, audio, video, and haptic inputs, performing complex ML inference and presenting outputs in any modality. It supports a wide range of frameworks (e.g., TensorFlow, PyTorch, ONNX) and can compile over 250 models, providing customers with an effortless experience and world-class performance-per-watt results. Complementing the hardware, SiMa.ai is designed for complete ML stack application development. It supports any ML workflow customers plan to deploy on the edge without compromising performance and ease of use. ...
  • 22
    Qualcomm Cloud AI SDK
    The Qualcomm Cloud AI SDK is a comprehensive software suite designed to optimize trained deep learning models for high-performance inference on Qualcomm Cloud AI 100 accelerators. It supports a wide range of AI frameworks, including TensorFlow, PyTorch, and ONNX, enabling developers to compile, optimize, and execute models efficiently. The SDK provides tools for model onboarding, tuning, and deployment, facilitating end-to-end workflows from model preparation to production deployment. Additionally, it offers resources such as model recipes, tutorials, and code samples to assist developers in accelerating AI development. ...
  • 23
    LiteRT

    LiteRT

    Google

    LiteRT (Lite Runtime), formerly known as TensorFlow Lite, is Google's high-performance runtime for on-device AI. It enables developers to deploy machine learning models across various platforms and microcontrollers. LiteRT supports models from TensorFlow, PyTorch, and JAX, converting them into the efficient FlatBuffers format (.tflite) for optimized on-device inference. Key features include low latency, enhanced privacy by processing data locally, reduced model and binary sizes, and efficient power consumption. The runtime offers SDKs in multiple languages such as Java/Kotlin, Swift, Objective-C, C++, and Python, facilitating integration into diverse applications. ...
    Starting Price: Free
  • 24
    Azure Data Science Virtual Machines
    ...Examples, templates and sample notebooks built or tested by Microsoft are provided on the VMs to enable easy onboarding to the various tools and capabilities such as Neural Networks (PYTorch, Tensorflow, etc.), Data Wrangling, R, Python, Julia, and SQL Server.
    Starting Price: $0.005
  • 25
    Ray

    Ray

    Anyscale

    ...Easily scale compute-heavy machine learning workloads like deep learning, model serving, and hyperparameter tuning with a strong ecosystem of distributed libraries. Scale existing workloads (for eg. Pytorch) on Ray with minimal effort by tapping into integrations. Native Ray libraries, such as Ray Tune and Ray Serve, lower the effort to scale the most compute-intensive machine learning workloads, such as hyperparameter tuning, training deep learning models, and reinforcement learning. For example, get started with distributed hyperparameter tuning in just 10 lines of code. ...
    Starting Price: Free
  • 26
    IBM Distributed AI APIs
    ...The Distributed AI APIs do not focus on the basic requirements of creating and deploying AI pipelines, for example, model training and model serving. You would use your favorite open-source packages such as TensorFlow or PyTorch. Then, you can containerize your application, including the AI pipeline, and deploy these containers at the distributed locations. In many cases, it’s useful to use a container orchestrator such as Kubernetes or OpenShift operators to automate the deployment process.
  • 27
    AWS Deep Learning AMIs
    AWS Deep Learning AMIs (DLAMI) provides ML practitioners and researchers with a curated and secure set of frameworks, dependencies, and tools to accelerate deep learning in the cloud. Built for Amazon Linux and Ubuntu, Amazon Machine Images (AMIs) come preconfigured with TensorFlow, PyTorch, Apache MXNet, Chainer, Microsoft Cognitive Toolkit (CNTK), Gluon, Horovod, and Keras, allowing you to quickly deploy and run these frameworks and tools at scale. Develop advanced ML models at scale to develop autonomous vehicle (AV) technology safely by validating models with millions of supported virtual tests. Accelerate the installation and configuration of AWS instances, and speed up experimentation and evaluation with up-to-date frameworks and libraries, including Hugging Face Transformers. ...
  • 28
    SuperDuperDB

    SuperDuperDB

    SuperDuperDB

    ...No need to introduce an additional database and duplicate your data to use vector search and build on top of it. SuperDuperDB enables vector search in your existing database. Integrate and combine models from Sklearn, PyTorch, and HuggingFace with AI APIs such as OpenAI to build even the most complex AI applications and workflows. Deploy all your AI models to automatically compute outputs (inference) in your datastore in a single environment with simple Python commands.
  • 29
    Gemma 2

    Gemma 2

    Google

    ...Gemma models achieve exceptional comparative results in their 2B, 7B, 9B, and 27B sizes, even outperforming some larger open models. With Keras 3.0, enjoy seamless compatibility with JAX, TensorFlow, and PyTorch, allowing you to effortlessly choose and change frameworks based on task. Redesigned to deliver outstanding performance and unmatched efficiency, Gemma 2 is optimized for incredibly fast inference on various hardware. The Gemma family of models offers different models that are optimized for specific use cases and adapt to your needs. ...
  • 30
    Google AI Edge
    ...It supports cross-platform compatibility, allowing the same model to run seamlessly across embedded systems. It is also multi-framework compatible, working with models from JAX, Keras, PyTorch, and TensorFlow. Key components include low-code APIs for common AI tasks through MediaPipe, enabling quick integration of generative AI, vision, text, and audio functionalities. Visualize the transformation of your model through conversion and quantification. Overlays the results of the comparisons to debug the hotspots. Explore, debug, and compare your models visually. ...
    Starting Price: Free
  • Previous
  • You're on page 1
  • 2
  • Next