Showing 13 open source projects for "multiple sequence alignment"

View related business solutions
  • Data management solutions for confident marketing Icon
    Data management solutions for confident marketing

    For companies wanting a complete Data Management solution that is native to Salesforce

    Verify, deduplicate, manipulate, and assign records automatically to keep your CRM data accurate, complete, and ready for business.
    Learn More
  • Securden Privileged Account Manager Icon
    Securden Privileged Account Manager

    Unified Privileged Access Management

    Discover and manage administrator, service, and web app passwords, keys, and identities. Automate management with approval workflows. Centrally control, audit, monitor, and record all access to critical IT assets.
    Learn More
  • 1
    WhisperJAV

    WhisperJAV

    Uses Qwen3-ASR, local LLM, Whisper, TEN-VAD

    WhisperJAV is an open-source speech transcription pipeline designed specifically for generating subtitles for Japanese adult video content. The project addresses challenges that standard speech recognition models face when transcribing this type of audio, which often includes low signal-to-noise ratios and large numbers of non-verbal vocalizations. Traditional automatic speech recognition systems can misinterpret these sounds as words, leading to inaccurate transcripts. WhisperJAV introduces...
    Downloads: 20 This Week
    Last Update:
    See Project
  • 2
    PKU Beaver

    PKU Beaver

    Constrained Value Alignment via Safe Reinforcement Learning

    PKU Beaver is an open-source research project focused on improving the safety alignment of large language models through reinforcement learning from human feedback under explicit safety constraints. The framework introduces techniques that separate helpfulness and harmlessness signals during training, allowing models to optimize for useful responses while minimizing harmful behavior. To support this process, the project provides datasets containing human-labeled examples that encode both performance preferences and safety constraints across multiple dimensions. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 3
    Qwen3

    Qwen3

    Qwen3 is the large language model series developed by Qwen team

    ...The latest updated version, Qwen3-235B-A22B-Instruct-2507, features significant improvements in instruction-following, reasoning, knowledge coverage, and long-context understanding up to 256K tokens. It delivers higher quality and more helpful text generation across multiple languages and domains, including mathematics, coding, science, and tool usage. Various quantized versions, tools/pipelines provided for inference using quantized formats (e.g. GGUF, etc.). Coverage for many languages in training and usage, alignment with human preferences in open-ended tasks, etc.
    Downloads: 31 This Week
    Last Update:
    See Project
  • 4
    In-The-Wild Jailbreak Prompts on LLMs

    In-The-Wild Jailbreak Prompts on LLMs

    A dataset consists of 15,140 ChatGPT prompts from Reddit

    ...The dataset includes thousands of prompts collected across multiple platforms and represents one of the largest collections of jailbreak attempts available for research.
    Downloads: 1 This Week
    Last Update:
    See Project
  • SoftCo: Enterprise Invoice and P2P Automation Software Icon
    SoftCo: Enterprise Invoice and P2P Automation Software

    For companies that process over 20,000 invoices per year

    SoftCo Accounts Payable Automation processes all PO and non-PO supplier invoices electronically from capture and matching through to invoice approval and query management. SoftCoAP delivers unparalleled touchless automation by embedding AI across matching, coding, routing, and exception handling to minimize the number of supplier invoices requiring manual intervention. The result is 89% processing savings, supported by a context-aware AI Assistant that helps users understand exceptions, answer questions, and take the right action faster.
    Learn More
  • 5
    RLHF-Reward-Modeling

    RLHF-Reward-Modeling

    Recipes to train reward model for RLHF

    ...In RLHF pipelines, reward models are responsible for evaluating generated responses and assigning scores that guide the model toward outputs that better match human preferences. The repository provides training recipes and implementations for building reward and preference models using modern machine learning frameworks. It supports multiple optimization strategies commonly used in alignment pipelines, including reinforcement learning with PPO, iterative supervised fine-tuning using rejection sampling, and direct preference optimization methods. The project also includes evaluation results showing that the trained reward models can achieve competitive performance compared with other open-source alignment systems.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 6
    LLM Course

    LLM Course

    Course to get into Large Language Models (LLMs)

    ...It emphasizes reproducible experiments: each step is demonstrated with runnable code, clear dependencies, and references to commonly used open-source models and libraries. Learners get exposure to multiple adaptation strategies—LoRA/QLoRA, instruction fine-tuning, and alignment techniques—so they can choose approaches that fit their hardware and budgets. The materials also cover inference optimization and quantization to make serving LLMs feasible on commodity GPUs or even CPUs, which is crucial for side projects and startups. Evaluation is treated as a first-class topic, with examples of automatic and human-in-the-loop methods to catch regressions and verify quality beyond simple loss values. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 7
    Hephaestus

    Hephaestus

    Semi-Structured Agentic Framework. Workflows build themselves

    Hephaestus is an open-source semi-structured agentic framework designed to orchestrate multiple AI agents working together on complex tasks. Instead of relying entirely on predefined workflows, the framework allows agents to dynamically create tasks as they explore a problem space. Developers define high-level phases such as analysis, implementation, and testing, while agents generate specific subtasks within those phases. The system continuously monitors agent behavior and task progression,...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 8
    LLMSurvey

    LLMSurvey

    A Survey of Large Language Models

    ...The repository organizes hundreds of research papers into thematic sections that reflect the main areas of LLM research, including model architectures, training strategies, evaluation benchmarks, alignment techniques, and downstream applications. By structuring the literature in a navigable format, LLMSurvey allows researchers and practitioners to quickly explore important publications in the field without manually searching through multiple databases.
    Downloads: 0 This Week
    Last Update:
    See Project
  • 9
    CodeGen

    CodeGen

    Open-source model for program synthesis

    CodeGen is a family of open-source large language models designed specifically for program synthesis and code generation tasks. Developed by Salesforce Research, the models are trained on large datasets containing both natural language and programming language content. This allows them to translate natural language descriptions into functional code across a variety of programming languages. CodeGen supports multi-turn program synthesis, meaning it can generate complex programs through a...
    Downloads: 1 This Week
    Last Update:
    See Project
  • The AI workplace management platform Icon
    The AI workplace management platform

    Plan smart spaces, connect teams, manage assets, and get insights with the leading AI-powered operating system for the built world.

    By combining AI workflows, predictive intelligence, and automated insights, OfficeSpace gives leaders a complete view of how their spaces are used and how people work. Facilities, IT, HR, and Real Estate teams use OfficeSpace to optimize space utilization, enhance employee experience, and reduce portfolio costs with precision.
    Learn More
  • 10
    Qwen2.5-Math

    Qwen2.5-Math

    A series of math-specific large language models of our Qwen2 series

    Qwen2.5-Math is a series of mathematics-specialized large language models in the Qwen2 family, released by Alibaba’s QwenLM. It includes base models (1.5B / 7B / 72B parameters), instruction-tuned versions, and a reward model (RM) to improve alignment. Unlike its predecessor Qwen2-Math, Qwen2.5-Math supports both Chain-of-Thought (CoT) reasoning and Tool-Integrated Reasoning (TIR) for solving math problems, and works in both Chinese and English. It is optimized for solving mathematical...
    Downloads: 1 This Week
    Last Update:
    See Project
  • 11
    dLLM

    dLLM

    dLLM: Simple Diffusion Language Modeling

    ...Unlike traditional autoregressive models that generate text sequentially token by token, diffusion language models generate text through an iterative denoising process that refines masked tokens over multiple steps. This approach allows models to reason over the entire sequence simultaneously and potentially produce more coherent outputs with bidirectional context. The project provides an integrated pipeline that standardizes how diffusion language models are trained, evaluated, and deployed, helping researchers reproduce experiments and compare results more easily. ...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 12
    InternLM

    InternLM

    Official release of InternLM series

    InternLM is an open-source family of multilingual foundation and chat models, accompanied by an ecosystem that supports training, inference, and application development. The repository highlights multiple model sizes intended to serve different needs, from efficient research and prototyping to more capable deployments for complex scenarios. Beyond model weights, the project emphasizes an ecosystem view, pointing developers to compatible tools and projects across training and inference so...
    Downloads: 0 This Week
    Last Update:
    See Project
  • 13
    LLMDataHub

    LLMDataHub

    Quick guide (especially) for trending instruction finetuning dataset

    ...The repository also highlights datasets suitable for reinforcement learning from human feedback and other alignment strategies used in modern language model training.
    Downloads: 0 This Week
    Last Update:
    See Project
  • Previous
  • You're on page 1
  • Next
MongoDB Logo MongoDB