elie.de hat dies direkt geteilt
IBM 💙 Open Source Our AI platform, watsonx, is powered by a stack of open source tech, enhancing AI workflows enterprise readiness. Here's the list of key projects: 𝗧𝗿𝗮𝗶𝗻𝗶𝗻𝗴 & 𝗩𝗮𝗹𝗶𝗱𝗮𝘁𝗶𝗼𝗻: - CodeFlare: Simplifies the scaling and management of distributed AI workloads by providing an easy-to-use interface for resource allocation, job submission, and workload management. - Ray / KubeRay: A framework for scaling distributed Python workloads. KubeRay integrates Ray with Kubernetes, enabling distributed AI tasks to run efficiently across clusters. - PyTorch: An open-source framework for deep learning model development, supporting both small and large distributed training, ideal for building AI models with over 10 billion parameters. - Kubeflow Training Operator: Orchestrates distributed training jobs across Kubernetes, supporting popular ML frameworks like PyTorch and TensorFlow for scalable AI model training. - Job Scheduler (Kueue/MCAD): Manages job scheduling and resource quotas, ensuring that distributed AI workloads are only started when sufficient resources are available. 𝗧𝘂𝗻𝗶𝗻𝗴 & 𝗜𝗻𝗳𝗲𝗿𝗲𝗻𝗰𝗲: - KServe: A Kubernetes-based platform for serving machine learning models at scale, providing production-level model inference for frameworks. - fms-hf-tuning: A collection of recipes for fine-tuning Hugging Face models using PyTorch’s distributed APIs, optimized for performance and scalability. - vLLM: A fast and flexible library designed for serving LLMs in both batch and real-time scenarios. - TGIS (Text Generation Inference Server): IBM’s fork of Hugging Face’s TGI, optimized for serving LLMs with high performance. - PyTorch: Used for both training and inference, this is a core framework in watsonx. - Hugging Face libraries: Offers a rich collection of pre-trained models and datasets, to provide cutting-edge AI capabilities. - Kubernetes DRA/InstaSlice: DRA allows for dynamic resource allocation in Kubernetes clusters, while InstaSlice facilitates resource sharing, particularly for GPU-intensive AI tasks. 𝗔𝗜 𝗗𝗲𝘃𝗲𝗹𝗼𝗽𝗺𝗲𝗻𝘁 𝗟𝗶𝗳𝗲𝗰𝘆𝗰𝗹𝗲: - Kubeflow & Pipelines: Provides end-to-end orchestration for AI workflows, automating everything from data preprocessing to model deployment and monitoring. - Open Data Hub: A comprehensive platform of tools for the entire AI lifecycle, from model development to deployment. - InstructLab: A project for shaping LLMs, allowing developers to enhance model capabilities by contributing skills and knowledge. - Granite models: IBM’s open source LLMs, spanning various modalities and trained on high-quality data. We're committed to the future of Open Source and its impact on the AI community.