latticearc

Projects that follow the best practices below can voluntarily self-certify and show that they've achieved an Open Source Security Foundation (OpenSSF) best practices badge.

There is no set of practices that can guarantee that software will never have defects or vulnerabilities; even formal methods can fail if the specifications or assumptions are wrong. Nor is there any set of practices that can guarantee that a project will sustain a healthy and well-functioning development community. However, following best practices can help improve the results of projects. For example, some practices enable multi-person review before release, which can both help find otherwise hard-to-find technical vulnerabilities and help build trust and a desire for repeated interaction among developers from different companies. To earn a badge, all MUST and MUST NOT criteria must be met, all SHOULD criteria must be met OR be unmet with justification, and all SUGGESTED criteria must be met OR unmet (we want them considered at least). If you want to enter justification text as a generic comment, instead of being a rationale that the situation is acceptable, start the text block with '//' followed by a space. Feedback is welcome via the GitHub site as issues or pull requests There is also a mailing list for general discussion.

We gladly provide the information in several locales, however, if there is any conflict or inconsistency between the translations, the English version is the authoritative version.
If this is your project, please show your badge status on your project page! The badge status looks like this: Badge level for project 11857 is passing Here is how to embed it:
You can show your badge status by embedding this in your markdown file:
[![OpenSSF Best Practices](https://www.bestpractices.dev/projects/11857/badge)](https://www.bestpractices.dev/projects/11857)
or by embedding this in your HTML:
<a href="https://www.bestpractices.dev/projects/11857"><img src="https://www.bestpractices.dev/projects/11857/badge"></a>


These are the Passing level criteria. You can also view the Silver or Gold level criteria.

Baseline Series: Baseline Level 1 Baseline Level 2 Baseline Level 3

        

 Basics 13/13

  • General

    Note that other projects may use the same name.

    LatticeArc is a post-quantum cryptography library for Rust implementing NIST FIPS 203-206 standards (ML-KEM, ML-DSA, SLH-DSA, FN-DSA) for quantum-resistant encryption and digital signatures.

    Please use SPDX license expression format; examples include "Apache-2.0", "BSD-2-Clause", "BSD-3-Clause", "GPL-2.0+", "LGPL-3.0+", "MIT", and "(BSD-2-Clause OR Ruby)". Do not include single quotes or double quotes.
    If there is more than one language, list them as comma-separated values (spaces optional) and sort them from most to least used. If there is a long list, please list at least the first three most common ones. If there is no language (e.g., this is a documentation-only or test-only project), use the single character "-". Please use a conventional capitalization for each language, e.g., "JavaScript".
    The Common Platform Enumeration (CPE) is a structured naming scheme for information technology systems, software, and packages. It is used in a number of systems and databases when reporting vulnerabilities.
  • Basic project website content


    The project website MUST succinctly describe what the software does (what problem does it solve?). [description_good]
    This MUST be in language that potential users can understand (e.g., it uses minimal jargon).

    The README.md clearly describes the problem and solution:

    Problem: "Current public-key cryptography (RSA, ECC) will be broken by quantum computers running Shor's algorithm. While large-scale quantum computers don't exist yet,
    encrypted data captured today can be decrypted in the future—a threat known as 'harvest now, decrypt later.'"

    Solution: "LatticeArc is a post-quantum cryptography library for Rust, implementing the NIST FIPS 203-206 standards for quantum-resistant encryption and digital
    signatures."

    Key features are succinctly listed:

    • ML-KEM (FIPS 203) - Key encapsulation
    • ML-DSA (FIPS 204) - Digital signatures
    • SLH-DSA (FIPS 205) - Hash-based signatures
    • FN-DSA (FIPS 206) - Lattice-based signatures

    The README also explains why hybrid mode is used (defense in depth) with clear diagrams.

    URL: https://github.com/LatticeArc/latticearc/blob/main/README.md



    The project website MUST provide information on how to: obtain, provide feedback (as bug reports or enhancements), and contribute to the software. [interact]

    The project provides clear information for all three requirements:

    1. OBTAIN:

      • README.md shows: cargo dependency latticearc = "0.1"
      • Crates.io badge links to package
      • GitHub repository for source code
    2. PROVIDE FEEDBACK:

      • GitHub Issues for bug reports and feature requests
      • SECURITY.md for vulnerability reporting (security@latticearc.com)
      • GitHub Discussions (if enabled)
    3. CONTRIBUTE:

      • CONTRIBUTING.md with:
        • Development setup instructions
        • Code style requirements (rustfmt, clippy)
        • Testing requirements
        • Pull request process
        • Code of conduct reference

    URLs:



    The information on how to contribute MUST explain the contribution process (e.g., are pull requests used?) (URL required) [contribution]
    We presume that projects on GitHub use issues and pull requests unless otherwise noted. This information can be short, e.g., stating that the project uses pull requests, an issue tracker, or posts to a mailing list (which one?)

    CONTRIBUTING.md clearly explains the contribution process:

    1. Pull Request Process (Section "Submitting Changes"):

      • Create PR against main branch
      • Fill out PR template
      • Ensure CI passes
      • Request review from maintainers
      • Address feedback
      • PRs require 2 approvals
    2. Branch Strategy:

      • feature/description for new features
      • fix/description for bug fixes
      • docs/description for documentation
    3. Commit Convention:

      • Conventional Commits format required
      • Types: feat, fix, docs, test, refactor, perf, chore, security
    4. PR Checklist provided:

      • Code compiles without warnings
      • All tests pass
      • New code has tests
      • Documentation updated
      • CHANGELOG.md updated

    URL: https://github.com/LatticeArc/latticearc/blob/main/CONTRIBUTING.md



    The information on how to contribute SHOULD include the requirements for acceptable contributions (e.g., a reference to any required coding standard). (URL required) [contribution_requirements]

    CONTRIBUTING.md specifies comprehensive requirements for acceptable contributions:

    1. Code Style Requirements:

      • rustfmt with project config (rustfmt.toml)
      • Max line width: 100 characters
      • Must pass: cargo clippy --workspace --all-targets --all-features -- -D warnings
    2. Testing Requirements:

      • Unit tests: 90%+ line coverage
      • Overall: 80%+ coverage
      • All public APIs must have tests
      • Must pass: cargo test --workspace --all-features
    3. Documentation Requirements:

      • All public items must be documented
      • Include # Examples section for complex APIs
      • Include # Errors section for fallible functions
      • Include # Panics section if applicable
    4. Security Requirements:

      • No unsafe code in cryptographic paths
      • Use constant-time operations for secret data
      • Zeroize sensitive data on drop
      • Validate all inputs
    5. Prohibited Patterns:

      • No unwrap()/expect() in library code
      • No timing-leak comparisons
      • No unzeroized sensitive data
      • No unsafe code
    6. Required Patterns:

      • Proper error handling with ?
      • Constant-time comparison (subtle::ConstantTimeEq)
      • Automatic zeroization (zeroize crate)

    URL: https://github.com/LatticeArc/latticearc/blob/main/CONTRIBUTING.md


  • FLOSS license


    The software produced by the project MUST be released as FLOSS. [floss_license]
    FLOSS is software released in a way that meets the Open Source Definition or Free Software Definition. Examples of such licenses include the CC0, MIT, BSD 2-clause, BSD 3-clause revised, Apache 2.0, Lesser GNU General Public License (LGPL), and the GNU General Public License (GPL). For our purposes, this means that the license MUST be: The software MAY also be licensed other ways (e.g., "GPLv2 or proprietary" is acceptable).

    LatticeArc is released under the Apache License 2.0, an OSI-approved free/libre open source software license.

    • LICENSE file contains full Apache 2.0 text
    • Cargo.toml declares: license = "Apache-2.0"
    • README.md displays license badge
    • All source files are covered under this license

    Apache 2.0 is approved by:

    • Open Source Initiative (OSI)
    • Free Software Foundation (as a free software license)
    • SPDX identifier: Apache-2.0

    URL: https://github.com/LatticeArc/latticearc/blob/main/LICENSE The Apache-2.0 license is approved by the Open Source Initiative (OSI).



    It is SUGGESTED that any required license(s) for the software produced by the project be approved by the Open Source Initiative (OSI). [floss_license_osi]
    The OSI uses a rigorous approval process to determine which licenses are OSS.

    Apache License 2.0 is approved by the Open Source Initiative (OSI).

    OSI Approval: https://opensource.org/licenses/Apache-2.0

    Apache 2.0 is one of the most widely used OSI-approved licenses, used by:

    • Apache Software Foundation projects
    • Google (Android, TensorFlow, Kubernetes)
    • Many Rust ecosystem crates

    The license allows:

    • Commercial use
    • Modification
    • Distribution
    • Patent use
    • Private use

    With conditions:

    • License and copyright notice
    • State changes

    URL:



    The project MUST post the license(s) of its results in a standard location in their source repository. (URL required) [license_location]
    One convention is posting the license as a top-level file named LICENSE or COPYING, which MAY be followed by an extension such as ".txt" or ".md". An alternative convention is to have a directory named LICENSES containing license file(s); these files are typically named as their SPDX license identifier followed by an appropriate file extension, as described in the REUSE Specification. Note that this criterion is only a requirement on the source repository. You do NOT need to include the license file when generating something from the source code (such as an executable, package, or container). For example, when generating an R package for the Comprehensive R Archive Network (CRAN), follow standard CRAN practice: if the license is a standard license, use the standard short license specification (to avoid installing yet another copy of the text) and list the LICENSE file in an exclusion file such as .Rbuildignore. Similarly, when creating a Debian package, you may put a link in the copyright file to the license text in /usr/share/common-licenses, and exclude the license file from the created package (e.g., by deleting the file after calling dh_auto_install). We encourage including machine-readable license information in generated formats where practical.

    The license is posted in the standard location:

    1. LICENSE file in repository root

      • Full Apache 2.0 license text
      • Standard filename recognized by GitHub, crates.io, and other tools
    2. Cargo.toml metadata

      • license = "Apache-2.0" in workspace package config
      • SPDX identifier for automated detection
    3. README.md badge

      • License badge visible at top of README
      • Links to LICENSE file

    GitHub automatically detects and displays the license in the repository sidebar.

    URL: https://github.com/LatticeArc/latticearc/blob/main/LICENSE


  • Documentation


    The project MUST provide basic documentation for the software produced by the project. [documentation_basics]
    This documentation must be in some media (such as text or video) that includes: how to install it, how to start it, how to use it (possibly with a tutorial using examples), and how to use it securely (e.g., what to do and what not to do) if that is an appropriate topic for the software. The security documentation need not be long. The project MAY use hypertext links to non-project material as documentation. If the project does not produce software, choose "not applicable" (N/A).

    The project provides comprehensive documentation:

    1. README.md - Project overview and quick start

      • Why post-quantum cryptography
      • Why hybrid mode
      • Quick start code examples
      • Algorithm selection tables
      • Crate structure overview
    2. docs/ folder with detailed guides:

      • UNIFIED_API_GUIDE.md - Complete API usage guide
      • SECURITY_GUIDE.md - Security best practices
      • NIST_COMPLIANCE.md - Standards compliance details
      • DESIGN.md - Architecture and design decisions
      • DEPLOYMENT_GUIDE.md - Production deployment
      • FAQ.md - Frequently asked questions
    3. API documentation (docs.rs)

      • Auto-generated from doc comments
      • All public types and functions documented
      • Working code examples
    4. CONTRIBUTING.md - Development documentation

    5. SECURITY.md - Security policy and reporting

    6. CHANGELOG.md - Version history

    URLs:



    The project MUST provide reference documentation that describes the external interface (both input and output) of the software produced by the project. [documentation_interface]
    The documentation of an external interface explains to an end-user or developer how to use it. This would include its application program interface (API) if the software has one. If it is a library, document the major classes/types and methods/functions that can be called. If it is a web application, define its URL interface (often its REST interface). If it is a command-line interface, document the parameters and options it supports. In many cases it's best if most of this documentation is automatically generated, so that this documentation stays synchronized with the software as it changes, but this isn't required. The project MAY use hypertext links to non-project material as documentation. Documentation MAY be automatically generated (where practical this is often the best way to do so). Documentation of a REST interface may be generated using Swagger/OpenAPI. Code interface documentation MAY be generated using tools such as JSDoc (JavaScript), ESDoc (JavaScript), pydoc (Python), devtools (R), pkgdown (R), and Doxygen (many). Merely having comments in implementation code is not sufficient to satisfy this criterion; there needs to be an easy way to see the information without reading through all the source code. If the project does not produce software, choose "not applicable" (N/A).

    The project provides reference documentation for all external interfaces:

    1. docs.rs API Reference (auto-generated from source):

      • All public functions with signatures
      • Parameter types and descriptions
      • Return types (EncryptedData, SignedData, Result<T>)
      • Error types and conditions
    2. Key Input/Output Documentation:

      Inputs:

      • CryptoConfig: SecurityLevel, UseCase, VerifiedSession
      • Key types: &[u8] for symmetric, PublicKey/PrivateKey for asymmetric
      • Data: &[u8] plaintext/message

      Outputs:

      • EncryptedData: { data, metadata, scheme, timestamp }
      • SignedData: { message, signature, public_key, scheme, timestamp }
      • Result<T, CoreError> with typed errors
    3. docs/UNIFIED_API_GUIDE.md:

      • Complete CryptoConfig builder documentation
      • SecurityLevel enum options and algorithm mappings
      • UseCase enum options and algorithm mappings
      • EncryptedData and SignedData structure reference
    4. Inline doc comments with # Arguments, # Returns, # Errors sections

    URLs:


  • Other


    The project sites (website, repository, and download URLs) MUST support HTTPS using TLS. [sites_https]
    This requires that the project home page URL and the version control repository URL begin with "https:", not "http:". You can get free certificates from Let's Encrypt. Projects MAY implement this criterion using (for example) GitHub pages, GitLab pages, or SourceForge project pages. If you support HTTP, we urge you to redirect the HTTP traffic to HTTPS.

    All project sites support HTTPS with TLS:

    1. Repository (GitHub):
      https://github.com/LatticeArc/latticearc

      • GitHub enforces HTTPS by default
      • HTTP redirects to HTTPS
      • TLS 1.2+ supported
    2. Package Registry (crates.io):
      https://crates.io/crates/latticearc

      • crates.io enforces HTTPS
      • All downloads over HTTPS
    3. Documentation (docs.rs):
      https://docs.rs/latticearc

      • docs.rs enforces HTTPS
    4. All URLs in README use HTTPS:

      • Badge URLs
      • Documentation links
      • External references

    URLs:



    The project MUST have one or more mechanisms for discussion (including proposed changes and issues) that are searchable, allow messages and topics to be addressed by URL, enable new people to participate in some of the discussions, and do not require client-side installation of proprietary software. [discussion]
    Examples of acceptable mechanisms include archived mailing list(s), GitHub issue and pull request discussions, Bugzilla, Mantis, and Trac. Asynchronous discussion mechanisms (like IRC) are acceptable if they meet these criteria; make sure there is a URL-addressable archiving mechanism. Proprietary JavaScript, while discouraged, is permitted.

    The project uses GitHub for discussions, which meets all requirements:

    1. GitHub Issues:

      • Searchable: Full-text search across all issues
      • URL addressable: Each issue has unique URL (e.g., /issues/123)
      • Open participation: Anyone with GitHub account can participate
      • Web-based: No proprietary client required
    2. GitHub Pull Requests:

      • Searchable: Full-text search
      • URL addressable: Each PR has unique URL
      • Code review comments addressable by URL
      • Open for public comment
    3. GitHub Discussions (if enabled):

      • Q&A, announcements, general discussion
      • Searchable and URL addressable
      • Threaded conversations

    All mechanisms:

    • Work in any web browser
    • Require no proprietary software
    • Allow anonymous viewing
    • Enable participation with free GitHub account

    URL: https://github.com/LatticeArc/latticearc/issues



    The project SHOULD provide documentation in English and be able to accept bug reports and comments about code in English. [english]
    English is currently the lingua franca of computer technology; supporting English increases the number of different potential developers and reviewers worldwide. A project can meet this criterion even if its core developers' primary language is not English.

    All project documentation and communication is in English:

    1. Documentation in English:

      • README.md
      • All docs/*.md files
      • CONTRIBUTING.md
      • SECURITY.md
      • CHANGELOG.md
      • Code comments and doc strings
      • API documentation (docs.rs)
    2. Bug Reports in English:

      • GitHub Issues accepts English submissions
      • Issue templates in English
      • Maintainers respond in English
    3. Code in English:

      • Variable/function names in English
      • Error messages in English
      • Log messages in English

    URL: https://github.com/LatticeArc/latticearc



    The project MUST be maintained. [maintained]
    As a minimum, the project should attempt to respond to significant problem and vulnerability reports. A project that is actively pursuing a badge is probably maintained. All projects and people have limited resources, and typical projects must reject some proposed changes, so limited resources and proposal rejections do not by themselves indicate an unmaintained project.

    When a project knows that it will no longer be maintained, it should set this criterion to "Unmet" and use the appropriate mechanism(s) to indicate to others that it is not being maintained. For example, use “DEPRECATED” as the first heading of its README, add “DEPRECATED” near the beginning of its home page, add “DEPRECATED” to the beginning of its code repository project description, add a no-maintenance-intended badge in its README and/or home page, mark it as deprecated in any package repositories (e.g., npm deprecate), and/or use the code repository's marking system to archive it (e.g., GitHub's "archive" setting, GitLab’s "archived" marking, Gerrit's "readonly" status, or SourceForge’s "abandoned" project status). Additional discussion can be found here.

    The project is actively maintained:

    1. Recent Activity:

      • 52 commits in the last 30 days
      • Multiple commits daily
      • Active development on main branch
    2. Maintenance Activities:

      • Bug fixes (hybrid signature verification)
      • Security updates (CI hardening, Docker image pinning)
      • Documentation updates
      • Code cleanup (11,500 lines dead code removed)
      • Test improvements
      • Version releases (v0.1.0 → v0.1.1 → v0.1.2)
    3. Responsiveness:

      • Issues addressed promptly
      • Security policy in place (SECURITY.md)
      • Active CI/CD pipeline
    4. Future Roadmap:

      • Codebase audit plan documented
      • Ongoing improvements planned

    URL: https://github.com/LatticeArc/latticearc/commits/main


 Change Control 9/9

  • Public version-controlled source repository


    The project MUST have a version-controlled source repository that is publicly readable and has a URL. [repo_public]
    The URL MAY be the same as the project URL. The project MAY use private (non-public) branches in specific cases while the change is not publicly released (e.g., for fixing a vulnerability before it is revealed to the public).

    The project uses a public GitHub repository with Git version control:

    1. Version Control System: Git

      • Full commit history preserved
      • Branching and tagging support
      • Distributed version control
    2. Publicly Readable:

      • No authentication required to view
      • Source code browsable via web
      • Clone/download available to anyone
    3. Repository URL:

    4. Features:

      • Commit history visible
      • Blame/annotation available
      • Diff viewing
      • Branch/tag browsing

    URL: https://github.com/LatticeArc/latticearc



    The project's source repository MUST track what changes were made, who made the changes, and when the changes were made. [repo_track]

    Git version control automatically tracks all required information:

    1. What Changes Were Made:

      • Full diff for every commit
      • Line-by-line change tracking
      • File additions, modifications, deletions recorded
    2. Who Made the Changes:

      • Author name and email on every commit
      • Committer information preserved
      • GitHub links commits to user profiles
    3. When Changes Were Made:

      • Timestamp on every commit
      • Author date and commit date recorded
      • Full chronological history

    Example commit metadata:
    commit 32e85d4
    Author: Kalyan Amaresam <...>
    Date: Fri Jan 30 23:22:15 2026 -0500

     docs: Add comprehensive codebase audit plan                                                                                                                         
    

    Viewable via:
    - git log (command line)
    - GitHub web interface (commits page)
    - git blame (per-line attribution)

    URL: https://github.com/LatticeArc/latticearc/commits/main Repository on GitHub, which uses git. git can track the changes, who made them, and when they were made.



    To enable collaborative review, the project's source repository MUST include interim versions for review between releases; it MUST NOT include only final releases. [repo_interim]
    Projects MAY choose to omit specific interim versions from their public source repositories (e.g., ones that fix specific non-public security vulnerabilities, may never be publicly released, or include material that cannot be legally posted and are not in the final release).

    The repository includes complete development history, not just releases:

    1. All Commits Visible:

      • Every individual commit is preserved
      • Development work between releases is tracked
      • 52+ commits in last 30 days (not just release tags)
    2. Interim Development:

      • Feature branches for work-in-progress
      • Individual commits for incremental changes
      • Pull requests show proposed changes before merge
    3. Release vs Development:

      • v0.1.0 (Jan 29) → 20+ commits → v0.1.1 (Jan 30) → 10+ commits → v0.1.2 (Jan 30)
      • All intermediate commits available for review
    4. Example Interim Commits:

      • "test(arc-core): Add comprehensive unified API tests"
      • "ci: Pin Docker base image by hash"
      • "docs: Clarify Apache vs Enterprise feature scope"

      These are development commits, not releases.

    URL: https://github.com/LatticeArc/latticearc/commits/main



    It is SUGGESTED that common distributed version control software be used (e.g., git) for the project's source repository. [repo_distributed]
    Git is not specifically required and projects can use centralized version control software (such as subversion) with justification.

    The project uses Git, the most widely-used distributed version control system:

    1. Version Control: Git

      • Distributed architecture
      • Full history on every clone
      • Offline commit capability
      • Industry standard
    2. Hosting: GitHub

      • Built on Git
      • Web interface for collaboration
      • Pull request workflow
      • Issue tracking integration
    3. Benefits:

      • Anyone can fork and contribute
      • Full redundancy (every clone is a backup)
      • Branching/merging workflow
      • Wide tooling support

    Git market share: ~95% of developers use Git (Stack Overflow surveys)

    URL: https://github.com/LatticeArc/latticearc Repository on GitHub, which uses git. git is distributed.


  • Unique version numbering


    The project results MUST have a unique version identifier for each release intended to be used by users. [version_unique]
    This MAY be met in a variety of ways including a commit IDs (such as git commit id or mercurial changeset id) or a version number (including version numbers that use semantic versioning or date-based schemes like YYYYMMDD).

    The project uses Semantic Versioning (SemVer) with unique identifiers:

    1. Version Format: MAJOR.MINOR.PATCH

      • v0.1.0 - Initial release (Jan 29, 2026)
      • v0.1.1 - Bug fixes (Jan 30, 2026)
      • v0.1.2 - Dead code cleanup (Jan 30, 2026)
    2. Version Sources:

      • Cargo.toml: version = "0.1.2"
      • Git tags for releases
      • CHANGELOG.md documents each version
      • crates.io publishes with version
    3. Uniqueness Guaranteed:

      • SemVer ensures no duplicate versions
      • crates.io prevents version reuse
      • Git tags are immutable
    4. Workspace Version:

      • Single version in [workspace.package]
      • All crates share same version
      • Consistent across entire project

    URL:



    It is SUGGESTED that the Semantic Versioning (SemVer) or Calendar Versioning (CalVer) version numbering format be used for releases. It is SUGGESTED that those who use CalVer include a micro level value. [version_semver]
    Projects should generally prefer whatever format is expected by their users, e.g., because it is the normal format used by their ecosystem. Many ecosystems prefer SemVer, and SemVer is generally preferred for application programmer interfaces (APIs) and software development kits (SDKs). CalVer tends to be used by projects that are large, have an unusually large number of independently-developed dependencies, have a constantly-changing scope, or are time-sensitive. It is SUGGESTED that those who use CalVer include a micro level value, because including a micro level supports simultaneously-maintained branches whenever that becomes necessary. Other version numbering formats may be used as version numbers, including git commit IDs or mercurial changeset IDs, as long as they uniquely identify versions. However, some alternatives (such as git commit IDs) can cause problems as release identifiers, because users may not be able to easily determine if they are up-to-date. The version ID format may be unimportant for identifying software releases if all recipients only run the latest version (e.g., it is the code for a single website or internet service that is constantly updated via continuous delivery).


    It is SUGGESTED that projects identify each release within their version control system. For example, it is SUGGESTED that those using git identify each release using git tags. [version_tags]

    The project uses git tags to identify releases:

    1. Git Tags Used:

      • v0.1.1 - "Fix hybrid signature verification and SecurityLevel docs"
    2. Tag Format:

      • Prefix: v
      • SemVer version number
      • Annotated with release description
    3. Note:

      • v0.1.0 and v0.1.2 tags should be added for completeness
      • Tagging process should be part of release checklist

    Tags viewable at:
    - git tag -l (command line)
    - GitHub Releases page

    URL: https://github.com/LatticeArc/latticearc/tags


    Recommendation: Add missing tags for consistency:
    git tag -a v0.1.0 <commit-hash> -m "v0.1.0: Initial release"
    git tag -a v0.1.2 <commit-hash> -m "v0.1.2: Dead code cleanup"
    git push origin --tags


  • Release notes


    The project MUST provide, in each release, release notes that are a human-readable summary of major changes in that release to help users determine if they should upgrade and what the upgrade impact will be. The release notes MUST NOT be the raw output of a version control log (e.g., the "git log" command results are not release notes). Projects whose results are not intended for reuse in multiple locations (such as the software for a single website or service) AND employ continuous delivery MAY select "N/A". (URL required) [release_notes]
    The release notes MAY be implemented in a variety of ways. Many projects provide them in a file named "NEWS", "CHANGELOG", or "ChangeLog", optionally with extensions such as ".txt", ".md", or ".html". Historically the term "change log" meant a log of every change, but to meet these criteria what is needed is a human-readable summary. The release notes MAY instead be provided by version control system mechanisms such as the GitHub Releases workflow.

    The project provides human-readable release notes in CHANGELOG.md:

    1. Format: Keep a Changelog (https://keepachangelog.com/)

      • Not raw git log output
      • Organized by category: Added, Changed, Fixed, Removed
      • Human-written summaries
    2. Content for Each Release:

      • Version number and date
      • Summary of major changes
      • Impact on users
      • Migration notes when needed
    3. Example (v0.1.2):

      [0.1.2] - 2026-01-30

      Removed

      • Dead Code Cleanup: Removed ~11,500 lines of unreachable code

      Added

      • Unified API Tests: Comprehensive test coverage

      Changed

      • Documentation: Clarified Apache vs Enterprise scope

      Notes

      • Explains what was removed and why
      • No breaking changes for users
    4. Helps Users Decide:

      • Breaking changes highlighted
      • Security fixes noted
      • New features listed
      • Deprecations documented

    URL: https://github.com/LatticeArc/latticearc/blob/main/CHANGELOG.md



    The release notes MUST identify every publicly known run-time vulnerability fixed in this release that already had a CVE assignment or similar when the release was created. This criterion may be marked as not applicable (N/A) if users typically cannot practically update the software themselves (e.g., as is often true for kernel updates). This criterion applies only to the project results, not to its dependencies. If there are no release notes or there have been no publicly known vulnerabilities, choose N/A. [release_notes_vulns]
    This criterion helps users determine if a given update will fix a vulnerability that is publicly known, to help users make an informed decision about updating. If users typically cannot practically update the software themselves on their computers, but must instead depend on one or more intermediaries to perform the update (as is often the case for a kernel and low-level software that is intertwined with a kernel), the project may choose "not applicable" (N/A) instead, since this additional information will not be helpful to those users. Similarly, a project may choose N/A if all recipients only run the latest version (e.g., it is the code for a single website or internet service that is constantly updated via continuous delivery). This criterion only applies to the project results, not its dependencies. Listing the vulnerabilities of all transitive dependencies of a project becomes unwieldy as dependencies increase and vary, and is unnecessary since tools that examine and track dependencies can do this in a more scalable way.

    Not applicable - No publicly known vulnerabilities:

    1. Project Status:

      • Pre-1.0 software (v0.1.2)
      • Initial release: January 29, 2026
      • No CVEs have been filed against LatticeArc
    2. No Vulnerabilities to Report:

      • No CVE assignments for this project
      • No security advisories issued
      • No known run-time vulnerabilities
    3. Future Commitment:

      • SECURITY.md documents vulnerability reporting process
      • CHANGELOG.md will document CVE fixes when applicable
      • Security fixes will be categorized under "### Security" section
    4. Verification:

      • No entries in NVD for "latticearc"
      • No GitHub Security Advisories for this repo

    URL: https://github.com/LatticeArc/latticearc/blob/main/SECURITY.md


 Reporting 8/8

  • Bug-reporting process


    The project MUST provide a process for users to submit bug reports (e.g., using an issue tracker or a mailing list). (URL required) [report_process]

    The project uses GitHub Issues for bug reports:

    1. Issue Tracker:

      • GitHub Issues enabled on repository
      • Publicly accessible
      • No account fee required (free GitHub account)
    2. Features:

      • Bug report templates (if configured)
      • Labels for categorization (bug, enhancement, etc.)
      • Assignees and milestones
      • Search and filtering
    3. Process:

      • Users click "New Issue"
      • Select issue type (bug report)
      • Fill in description, steps to reproduce
      • Submit for maintainer review
    4. Documentation:

      • CONTRIBUTING.md references issue tracker
      • README links to issues
      • SECURITY.md for security vulnerabilities (separate process)

    URL: https://github.com/LatticeArc/latticearc/issues



    The project SHOULD use an issue tracker for tracking individual issues. [report_tracker]

    The project uses GitHub Issues as its issue tracker:

    1. Issue Tracker: GitHub Issues

      • Integrated with repository
      • Full-featured issue tracking
      • Free and publicly accessible
    2. Tracking Features:

      • Unique issue numbers (#1, #2, etc.)
      • Open/closed status
      • Labels (bug, enhancement, documentation, etc.)
      • Milestones for release planning
      • Assignees for ownership
      • Cross-references between issues and PRs
    3. Organization:

      • Search and filter capabilities
      • Sort by date, comments, reactions
      • Project boards for workflow (if enabled)
    4. Integration:

      • Links to commits that fix issues
      • Auto-close issues via commit messages
      • PR references to related issues

    URL: https://github.com/LatticeArc/latticearc/issues



    The project MUST acknowledge a majority of bug reports submitted in the last 2-12 months (inclusive); the response need not include a fix. [report_responses]

    The project acknowledges bug reports:

    1. Current Status:

      • Project is newly released (v0.1.0 on Jan 29, 2026)
      • No external bug reports submitted yet
      • No issues pending response
    2. Response Commitment:

      • Maintainers actively monitor issues
      • CONTRIBUTING.md encourages issue submission
      • Response expected within reasonable timeframe
    3. Evidence of Responsiveness:

      • Internal issues addressed promptly
      • Active development (52 commits in 30 days)
      • Same-day fixes for discovered issues (e.g., v0.1.1 hybrid signature fix)
    4. Process in Place:

      • GitHub Issues enabled
      • Notifications configured
      • Maintainers committed to engagement

    Note: As a new project, 100% of bug reports (0 of 0) have been acknowledged.

    URL: https://github.com/LatticeArc/latticearc/issues



    The project SHOULD respond to a majority (>50%) of enhancement requests in the last 2-12 months (inclusive). [enhancement_responses]
    The response MAY be 'no' or a discussion about its merits. The goal is simply that there be some response to some requests, which indicates that the project is still alive. For purposes of this criterion, projects need not count fake requests (e.g., from spammers or automated systems). If a project is no longer making enhancements, please select "unmet" and include the URL that makes this situation clear to users. If a project tends to be overwhelmed by the number of enhancement requests, please select "unmet" and explain.

    LatticeArc is a newly public project (v0.1.0 released 2026-01-29), so the 2-12 month window has minimal data. However:

    1. Active development - 3 releases in 2 days demonstrates high responsiveness
    2. GitHub Issues enabled - Public tracker for enhancement requests
    3. Responsive to feedback - CHANGELOG shows rapid iteration on user-facing improvements (SecurityLevel redesign, API fixes, documentation updates)


    The project MUST have a publicly available archive for reports and responses for later searching. (URL required) [report_archive]

    GitHub Issues provides a publicly accessible, searchable archive of all bug reports, enhancement requests, and their responses.
    URL: https://github.com/latticearc/latticearc/issues


  • Vulnerability report process


    The project MUST publish the process for reporting vulnerabilities on the project site. (URL required) [vulnerability_report_process]
    Projects hosted on GitHub SHOULD consider enabling privately reporting a security vulnerability. Projects on GitLab SHOULD consider using its ability for privately reporting a vulnerability. Projects MAY identify a mailing address on https://PROJECTSITE/security, often in the form security@example.org. This vulnerability reporting process MAY be the same as its bug reporting process. Vulnerability reports MAY always be public, but many projects have a private vulnerability reporting mechanism.

    SECURITY.md documents the full vulnerability reporting process including private email (Security@LatticeArc.com), GitHub Security Advisory, response timelines, and coordinated disclosure policy.
    URL: https://github.com/latticearc/latticearc/blob/main/SECURITY.md



    If private vulnerability reports are supported, the project MUST include how to send the information in a way that is kept private. (URL required) [vulnerability_report_private]
    Examples include a private defect report submitted on the web using HTTPS (TLS) or an email encrypted using OpenPGP. If vulnerability reports are always public (so there are never private vulnerability reports), choose "not applicable" (N/A).

    SECURITY.md explicitly directs reporters away from public issues and provides two private channels: email (Security@LatticeArc.com) and GitHub Security Advisories.
    URL: https://github.com/latticearc/latticearc/blob/main/SECURITY.md



    The project's initial response time for any vulnerability report received in the last 6 months MUST be less than or equal to 14 days. [vulnerability_report_response]
    If there have been no vulnerabilities reported in the last 6 months, choose "not applicable" (N/A).

    SECURITY.md commits to 24-hour initial acknowledgment, well under the 14-day requirement. Project is newly public (January 2026) with no vulnerability reports received yet, but the response policy is documented and maintainers are committed to meeting these timelines


 Quality 13/13

  • Working build system


    If the software produced by the project requires building for use, the project MUST provide a working build system that can automatically rebuild the software from source code. [build]
    A build system determines what actions need to occur to rebuild the software (and in what order), and then performs those steps. For example, it can invoke a compiler to compile the source code. If an executable is created from source code, it must be possible to modify the project's source code and then generate an updated executable with those modifications. If the software produced by the project depends on external libraries, the build system does not need to build those external libraries. If there is no need to build anything to use the software after its source code is modified, select "not applicable" (N/A).

    Uses Cargo (Rust's standard build system). Single command cargo build --workspace rebuilds entire project from source. Workspace Cargo.toml coordinates all crates, Cargo.lock ensures reproducible dependencies.



    It is SUGGESTED that common tools be used for building the software. [build_common_tools]
    For example, Maven, Ant, cmake, the autotools, make, rake (Ruby), or devtools (R).

    Uses Cargo, the official Rust build system included with every Rust installation. No custom build tooling required. Additional common Rust ecosystem tools (cargo-audit, cargo-deny, cargo-fuzz) used for security and quality.



    The project SHOULD be buildable using only FLOSS tools. [build_floss_tools]

    Entire build chain is FLOSS. Rust toolchain (rustc, cargo) is MIT/Apache-2.0. All dependencies verified FLOSS via deny.toml which explicitly allows only MIT, Apache-2.0, BSD, ISC, CC0, MPL-2.0 licenses. No proprietary tools required.


  • Automated test suite


    The project MUST use at least one automated test suite that is publicly released as FLOSS (this test suite may be maintained as a separate FLOSS project). The project MUST clearly show or document how to run the test suite(s) (e.g., via a continuous integration (CI) script or via documentation in files such as BUILD.md, README.md, or CONTRIBUTING.md). [test]
    The project MAY use multiple automated test suites (e.g., one that runs quickly, vs. another that is more thorough but requires special equipment). There are many test frameworks and test support systems available, including Selenium (web browser automation), Junit (JVM, Java), RUnit (R), testthat (R).

    Uses Cargo's built-in test framework (MIT/Apache-2.0) plus proptest for property-based testing. Test commands documented in README.md (cargo test --workspace --all-features). CI automatically runs full test suite on every pull request via GitHub Actions.



    A test suite SHOULD be invocable in a standard way for that language. [test_invocation]
    For example, "make check", "mvn test", or "rake test" (Ruby).

    Uses cargo test, the universal standard for Rust projects. No custom test harness or non-standard invocation required. All standard Cargo test flags and conventions work as expected.



    It is SUGGESTED that the test suite cover most (or ideally all) the code branches, input fields, and functionality. [test_most]

    Targets 80% code coverage. Uses multiple strategies: unit tests, integration tests, property-based testing (proptest), NIST CAVP validation vectors, and fuzz testing. All cryptographic primitives have roundtrip and negative tests. Coverage measured via cargo-llvm-cov



    It is SUGGESTED that the project implement continuous integration (where new or changed code is frequently integrated into a central code repository and automated tests are run on the result). [test_continuous_integration]

    GitHub Actions runs full CI on every push and pull request. Pipeline includes build, test, clippy linting, format checking, security audit (cargo-audit), and dependency verification (cargo-deny). Daily scheduled fuzzing. PRs cannot merge until all checks pass.


  • New functionality testing


    The project MUST have a general policy (formal or not) that as major new functionality is added to the software produced by the project, tests of that functionality should be added to an automated test suite. [test_policy]
    As long as a policy is in place, even by word of mouth, that says developers should add tests to the automated test suite for major new functionality, select "Met."

    CONTRIBUTING.md explicitly requires tests for new functionality: 'Each PR should include tests for new functionality' and 'All public APIs must have tests.' PR template has test checkboxes, and CI blocks merging until tests pass. Coverage targets: 90% unit, 80% overall.



    The project MUST have evidence that the test_policy for adding tests has been adhered to in the most recent major changes to the software produced by the project. [tests_are_added]
    Major functionality would typically be mentioned in the release notes. Perfection is not required, merely evidence that tests are typically being added in practice to the automated test suite when new major functionality is added to the software produced by the project.

    Recent commits demonstrate test policy adherence. Example: commit 00da0fd added 6 comprehensive tests for new unified API functionality. CHANGELOG documents test additions for each release. PR process enforces test requirements via template checklist and CI.



    It is SUGGESTED that this policy on adding tests (see test_policy) be documented in the instructions for change proposals. [tests_documented_added]
    However, even an informal rule is acceptable as long as the tests are being added in practice.

    Test policy is documented in CONTRIBUTING.md with a dedicated 'Testing Requirements' section specifying coverage thresholds and test categories. PR template includes testing checklist that contributors must complete before submitting changes.


  • Warning flags


    The project MUST enable one or more compiler warning flags, a "safe" language mode, or use a separate "linter" tool to look for code quality errors or common simple mistakes, if there is at least one FLOSS tool that can implement this criterion in the selected language. [warnings]
    Examples of compiler warning flags include gcc/clang "-Wall". Examples of a "safe" language mode include JavaScript "use strict" and perl5's "use warnings". A separate "linter" tool is simply a tool that examines the source code to look for code quality errors or common simple mistakes. These are typically enabled within the source code or build instructions.

    Extensive lint configuration in Cargo.toml: unsafe_code is forbidden, 80+ Clippy rules at deny/warn level. CI runs cargo clippy -- -D warnings blocking merge on
    any warning. Rust's memory-safe design plus #![forbid(unsafe_code)] enforces safe language mode.



    The project MUST address warnings. [warnings_fixed]
    These are the warnings identified by the implementation of the warnings criterion. The project should fix warnings or mark them in the source code as false positives. Ideally there would be no warnings, but a project MAY accept some warnings (typically less than 1 warning per 100 lines or less than 10 warnings).

    Zero warnings in current codebase (verified with cargo clippy -- -D warnings). CI treats warnings as errors, blocking PRs with warnings. Commit history shows
    active warning remediation (e.g., 'fix: Resolve clippy field_reassign_with_default'). Warnings are fixed promptly, not suppressed.



    It is SUGGESTED that projects be maximally strict with warnings in the software produced by the project, where practical. [warnings_strict]
    Some warnings cannot be effectively enabled on some projects. What is needed is evidence that the project is striving to enable warning flags where it can, so that errors are detected early.

    Maximally strict configuration for cryptographic code. Uses forbid (non-overridable) for unsafe_code, deny (compile error) for 40+ lint rules including panic,
    unwrap, indexing. Only allows warn level where crypto algorithms require specific behavior (e.g., modular arithmetic). 80+ Clippy rules configured beyond defaults.


 Security 16/16

  • Secure development knowledge


    The project MUST have at least one primary developer who knows how to design secure software. (See ‘details’ for the exact requirements.) [know_secure_design]
    This requires understanding the following design principles, including the 8 principles from Saltzer and Schroeder:
    • economy of mechanism (keep the design as simple and small as practical, e.g., by adopting sweeping simplifications)
    • fail-safe defaults (access decisions should deny by default, and projects' installation should be secure by default)
    • complete mediation (every access that might be limited must be checked for authority and be non-bypassable)
    • open design (security mechanisms should not depend on attacker ignorance of its design, but instead on more easily protected and changed information like keys and passwords)
    • separation of privilege (ideally, access to important objects should depend on more than one condition, so that defeating one protection system won't enable complete access. E.G., multi-factor authentication, such as requiring both a password and a hardware token, is stronger than single-factor authentication)
    • least privilege (processes should operate with the least privilege necessary)
    • least common mechanism (the design should minimize the mechanisms common to more than one user and depended on by all users, e.g., directories for temporary files)
    • psychological acceptability (the human interface must be designed for ease of use - designing for "least astonishment" can help)
    • limited attack surface (the attack surface - the set of the different points where an attacker can try to enter or extract data - should be limited)
    • input validation with allowlists (inputs should typically be checked to determine if they are valid before they are accepted; this validation should use allowlists (which only accept known-good values), not denylists (which attempt to list known-bad values)).
    A "primary developer" in a project is anyone who is familiar with the project's code base, is comfortable making changes to it, and is acknowledged as such by most other participants in the project. A primary developer would typically make a number of contributions over the past year (via code, documentation, or answering questions). Developers would typically be considered primary developers if they initiated the project (and have not left the project more than three years ago), have the option of receiving information on a private vulnerability reporting channel (if there is one), can accept commits on behalf of the project, or perform final releases of the project software. If there is only one developer, that individual is the primary developer. Many books and courses are available to help you understand how to develop more secure software and discuss design. For example, the Secure Software Development Fundamentals course is a free set of three courses that explain how to develop more secure software (it's free if you audit it; for an extra fee you can earn a certificate to prove you learned the material).

    Primary developers demonstrate secure design expertise through: cryptographic architecture following NIST FIPS 203-206, memory-safe Rust with forbidden unsafe code, constant-time operations for secrets, zeroization of sensitive data, comprehensive security documentation with threat model, and secure-by-default configuration.



    At least one of the project's primary developers MUST know of common kinds of errors that lead to vulnerabilities in this kind of software, as well as at least one method to counter or mitigate each of them. [know_common_errors]
    Examples (depending on the type of software) include SQL injection, OS injection, classic buffer overflow, cross-site scripting, missing authentication, and missing authorization. See the CWE/SANS top 25 or OWASP Top 10 for commonly used lists. Many books and courses are available to help you understand how to develop more secure software and discuss common implementation errors that lead to vulnerabilities. For example, the Secure Software Development Fundamentals course is a free set of three courses that explain how to develop more secure software (it's free if you audit it; for an extra fee you can earn a certificate to prove you learned the material).

    Developers demonstrate knowledge of cryptographic vulnerabilities: timing attacks (mitigated via subtle crate), memory leaks (zeroize crate), buffer overflows (safe indexing, Rust bounds), panics (deny lint + Result types), weak crypto (NIST FIPS only). SECURITY.md and CONTRIBUTING.md document prohibited patterns and required mitigations.


  • Use basic good cryptographic practices

    Note that some software does not need to use cryptographic mechanisms. If your project produces software that (1) includes, activates, or enables encryption functionality, and (2) might be released from the United States (US) to outside the US or to a non-US-citizen, you may be legally required to take a few extra steps. Typically this just involves sending an email. For more information, see the encryption section of Understanding Open Source Technology & US Export Controls.

    The software produced by the project MUST use, by default, only cryptographic protocols and algorithms that are publicly published and reviewed by experts (if cryptographic protocols and algorithms are used). [crypto_published]
    These cryptographic criteria do not always apply because some software has no need to directly use cryptographic capabilities.

    All cryptographic algorithms are NIST FIPS or IETF RFC standards: ML-KEM (FIPS 203), ML-DSA (FIPS 204), SLH-DSA (FIPS 205), AES-GCM (FIPS 197), SHA-2/3 (FIPS 180/202), Ed25519 (RFC 8032), HKDF (RFC 5869). No proprietary algorithms. All underwent extensive public cryptanalysis.



    If the software produced by the project is an application or library, and its primary purpose is not to implement cryptography, then it SHOULD only call on software specifically designed to implement cryptographic functions; it SHOULD NOT re-implement its own. [crypto_call]

    LatticeArc's primary purpose IS cryptography. However, it follows the spirit: no custom primitives are implemented. All algorithms use audited crates (aws-lc-rs FIPS 140-3 validated, RustCrypto audited). LatticeArc provides API unification and hybrid composition, not novel crypto algorithms.



    All functionality in the software produced by the project that depends on cryptography MUST be implementable using FLOSS. [crypto_floss]

    All cryptographic functionality uses FLOSS libraries: aws-lc-rs (Apache-2.0), fips204/205 (MIT/Apache-2.0), ed25519-dalek (BSD-3), RustCrypto crates (MIT/Apache-2.0). License compliance enforced via deny.toml. No proprietary crypto dependencies.



    The security mechanisms within the software produced by the project MUST use default keylengths that at least meet the NIST minimum requirements through the year 2030 (as stated in 2012). It MUST be possible to configure the software so that smaller keylengths are completely disabled. [crypto_keylength]
    These minimum bitlengths are: symmetric key 112, factoring modulus 2048, discrete logarithm key 224, discrete logarithmic group 2048, elliptic curve 224, and hash 224 (password hashing is not covered by this bitlength, more information on password hashing can be found in the crypto_password_storage criterion). See https://www.keylength.com for a comparison of keylength recommendations from various organizations. The software MAY allow smaller keylengths in some configurations (ideally it would not, since this allows downgrade attacks, but shorter keylengths are sometimes necessary for interoperability).

    Defaults exceed NIST 2030 requirements: AES-256 (vs 128 min), P-256 (vs P-224 min), SHA-256 (vs SHA-224 min). Default is 'High' security level (NIST Level 3 / 192-bit). Weak algorithms not offered: no AES-128, no P-192, no SHA-1. Post-quantum algorithms (ML-KEM, ML-DSA) meet NIST FIPS standards.



    The default security mechanisms within the software produced by the project MUST NOT depend on broken cryptographic algorithms (e.g., MD4, MD5, single DES, RC4, Dual_EC_DRBG), or use cipher modes that are inappropriate to the context, unless they are necessary to implement an interoperable protocol (where the protocol implemented is the most recent version of that standard broadly supported by the network ecosystem, that ecosystem requires the use of such an algorithm or mode, and that ecosystem does not offer any more secure alternative). The documentation MUST describe any relevant security risks and any known mitigations if these broken algorithms or modes are necessary for an interoperable protocol. [crypto_working]
    ECB mode is almost never appropriate because it reveals identical blocks within the ciphertext as demonstrated by the ECB penguin, and CTR mode is often inappropriate because it does not perform authentication and causes duplicates if the input state is repeated. In many cases it's best to choose a block cipher algorithm mode designed to combine secrecy and authentication, e.g., Galois/Counter Mode (GCM) and EAX. Projects MAY allow users to enable broken mechanisms (e.g., during configuration) where necessary for compatibility, but then users know they're doing it.

    No broken algorithms available: no MD4/MD5/SHA-1, no DES/3DES/RC4, no Dual_EC_DRBG. Only authenticated encryption modes (AES-GCM, ChaCha20-Poly1305) - no ECB or unauthenticated CBC. Fresh codebase with no legacy interoperability requirements. Only NIST FIPS and IETF RFC approved algorithms.



    The default security mechanisms within the software produced by the project SHOULD NOT depend on cryptographic algorithms or modes with known serious weaknesses (e.g., the SHA-1 cryptographic hash algorithm or the CBC mode in SSH). [crypto_weaknesses]
    Concerns about CBC mode in SSH are discussed in CERT: SSH CBC vulnerability.

    All algorithms are current standards without known weaknesses: AES-256-GCM, ChaCha20-Poly1305, ML-KEM/ML-DSA/SLH-DSA (NIST 2024), SHA-256/SHA-3, Ed25519, P-256. Default hybrid mode provides defense-in-depth. cargo-audit monitors for vulnerabilities in CI.



    The security mechanisms within the software produced by the project SHOULD implement perfect forward secrecy for key agreement protocols so a session key derived from a set of long-term keys cannot be compromised if one of the long-term keys is compromised in the future. [crypto_pfs]

    Perfect Forward Secrecy achieved through ephemeral key encapsulation. Each encryption uses fresh ML-KEM/ECDH encapsulation to derive session keys. Long-term keys are signing-only (identity). TLS integration via rustls defaults to PFS cipher suites. Compromise of signing keys cannot reveal past session keys.



    If the software produced by the project causes the storing of passwords for authentication of external users, the passwords MUST be stored as iterated hashes with a per-user salt by using a key stretching (iterated) algorithm (e.g., Argon2id, Bcrypt, Scrypt, or PBKDF2). See also OWASP Password Storage Cheat Sheet. [crypto_password_storage]
    This criterion applies only when the software is enforcing authentication of users using passwords for external users (aka inbound authentication), such as server-side web applications. It does not apply in cases where the software stores passwords for authenticating into other systems (aka outbound authentication, e.g., the software implements a client for some other system), since at least parts of that software must have often access to the unhashed password.

    LatticeArc is a cryptographic library that does not store user passwords. However, it provides PBKDF2 (in arc-primitives/kdf/pbkdf2.rs) for applications requiring password-based key derivation with salt and configurable iterations, following OWASP Password Storage Cheat Sheet recommendations.



    The security mechanisms within the software produced by the project MUST generate all cryptographic keys and nonces using a cryptographically secure random number generator, and MUST NOT do so using generators that are cryptographically insecure. [crypto_random]
    A cryptographically secure random number generator may be a hardware random number generator, or it may be a cryptographically secure pseudo-random number generator (CSPRNG) using an algorithm such as Hash_DRBG, HMAC_DRBG, CTR_DRBG, Yarrow, or Fortuna. Examples of calls to secure random number generators include Java's java.security.SecureRandom and JavaScript's window.crypto.getRandomValues. Examples of calls to insecure random number generators include Java's java.util.Random and JavaScript's Math.random.

    All keys and nonces generated via CSPRNG. Uses getrandom (OS entropy), rand with std_rng (ChaCha12), and aws-lc-rs (FIPS 140-3 validated DRBG). Dedicated Csprng type in arc-primitives. No insecure RNGs (no time-based seeds, no SmallRng for crypto). Fresh random nonces/IVs per operation.


  • Secured delivery against man-in-the-middle (MITM) attacks


    The project MUST use a delivery mechanism that counters MITM attacks. Using https or ssh+scp is acceptable. [delivery_mitm]
    An even stronger mechanism is releasing the software with digitally signed packages, since that mitigates attacks on the distribution system, but this only works if the users can be confident that the public keys for signatures are correct and if the users will actually check the signature.

    All delivery via HTTPS or SSH. GitHub repository accessed via HTTPS/SSH. Dependencies from crates.io over HTTPS with checksum verification. Cargo.lock pins dependency hashes. deny.toml restricts sources to crates.io only. No HTTP or insecure delivery mechanisms.



    A cryptographic hash (e.g., a sha1sum) MUST NOT be retrieved over http and used without checking for a cryptographic signature. [delivery_unsigned]
    These hashes can be modified in transit.

    No HTTP hash retrieval. All dependencies from crates.io over HTTPS with Cargo's built-in checksum verification. Cargo.lock contains cryptographic checksums for reproducible builds. deny.toml denies unknown registries/git sources. CI actions pinned by SHA hash, not tags.


  • Publicly known vulnerabilities fixed


    There MUST be no unpatched vulnerabilities of medium or higher severity that have been publicly known for more than 60 days. [vulnerabilities_fixed_60_days]
    The vulnerability must be patched and released by the project itself (patches may be developed elsewhere). A vulnerability becomes publicly known (for this purpose) once it has a CVE with publicly released non-paywalled information (reported, for example, in the National Vulnerability Database) or when the project has been informed and the information has been released to the public (possibly by the project). A vulnerability is considered medium or higher severity if its Common Vulnerability Scoring System (CVSS) base qualitative score is medium or higher. In CVSS versions 2.0 through 3.1, this is equivalent to a CVSS score of 4.0 or higher. Projects may use the CVSS score as published in a widely-used vulnerability database (such as the National Vulnerability Database) using the most-recent version of CVSS reported in that database. Projects may instead calculate the severity themselves using the latest version of CVSS at the time of the vulnerability disclosure, if the calculation inputs are publicly revealed once the vulnerability is publicly known. Note: this means that users might be left vulnerable to all attackers worldwide for up to 60 days. This criterion is often much easier to meet than what Google recommends in Rebooting responsible disclosure, because Google recommends that the 60-day period start when the project is notified even if the report is not public. Also note that this badge criterion, like other criteria, applies to the individual project. Some projects are part of larger umbrella organizations or larger projects, possibly in multiple layers, and many projects feed their results to other organizations and projects as part of a potentially-complex supply chain. An individual project often cannot control the rest, but an individual project can work to release a vulnerability patch in a timely way. Therefore, we focus solely on the individual project's response time. Once a patch is available from the individual project, others can determine how to deal with the patch (e.g., they can update to the newer version or they can apply just the patch as a cherry-picked solution).

    No medium or higher severity vulnerabilities. cargo audit shows only 'unmaintained' warnings for transitive dev dependencies (not security issues). cargo-deny advisories check passes. CI runs cargo audit on every PR. Dependabot enabled for automated security updates.



    Projects SHOULD fix all critical vulnerabilities rapidly after they are reported. [vulnerabilities_critical_fixed]

    No critical vulnerabilities reported to date. SECURITY.md commits to 7-day fix timeline for critical issues. Demonstrated rapid development velocity (3 releases in 2 days). CI/CD enables quick releases, cargo audit catches new advisories, Dependabot automates security updates. Security review required for crypto changes.


  • Other security issues


    The public repositories MUST NOT leak a valid private credential (e.g., a working password or private key) that is intended to limit public access. [no_leaked_credentials]
    A project MAY leak "sample" credentials for testing and unimportant databases, as long as they are not intended to limit public access.

    No credential files in repository (.env, .pem, .key excluded via .gitignore). Git history contains only code references (parameter names, type names), not actual secrets. GitHub Secret Scanning enabled. Test data uses synthetic/generated keys, not real credentials.


 Analysis 8/8

  • Static code analysis


    At least one static code analysis tool (beyond compiler warnings and "safe" language modes) MUST be applied to any proposed major production release of the software before its release, if there is at least one FLOSS tool that implements this criterion in the selected language. [static_analysis]
    A static code analysis tool examines the software code (as source code, intermediate code, or executable) without executing it with specific inputs. For purposes of this criterion, compiler warnings and "safe" language modes do not count as static code analysis tools (these typically avoid deep analysis because speed is vital). Some static analysis tools focus on detecting generic defects, others focus on finding specific kinds of defects (such as vulnerabilities), and some do a combination. Examples of such static code analysis tools include cppcheck (C, C++), clang static analyzer (C, C++), SpotBugs (Java), FindBugs (Java) (including FindSecurityBugs), PMD (Java), Brakeman (Ruby on Rails), lintr (R), goodpractice (R), Coverity Quality Analyzer, SonarQube, Codacy, and HP Enterprise Fortify Static Code Analyzer. Larger lists of tools can be found in places such as the Wikipedia list of tools for static code analysis, OWASP information on static code analysis, NIST list of source code security analyzers, and Wheeler's list of static analysis tools. If there are no FLOSS static analysis tools available for the implementation language(s) used, you may select 'N/A'.

    Multiple static analysis tools run before release: Clippy (500+ lint rules, 80+ custom configured), cargo-audit (security vulnerabilities), cargo-deny (dependencies), CodeQL (GitHub SAST). All run in CI on every PR and block merge on failure. All tools are FLOSS.



    It is SUGGESTED that at least one of the static analysis tools used for the static_analysis criterion include rules or approaches to look for common vulnerabilities in the analyzed language or environment. [static_analysis_common_vulnerabilities]
    Static analysis tools that are specifically designed to look for common vulnerabilities are more likely to find them. That said, using any static tools will typically help find some problems, so we are suggesting but not requiring this for the 'passing' level badge.

    Multiple tools check for common vulnerabilities: Clippy (panic/crash, buffer access, integer overflow), cargo-audit (900+ CVEs from RustSec database), CodeQL (injection, path traversal, crypto misuse). Security lints explicitly configured at deny level in Cargo.toml.



    All medium and higher severity exploitable vulnerabilities discovered with static code analysis MUST be fixed in a timely way after they are confirmed. [static_analysis_fixed]
    A vulnerability is considered medium or higher severity if its Common Vulnerability Scoring System (CVSS) base qualitative score is medium or higher. In CVSS versions 2.0 through 3.1, this is equivalent to a CVSS score of 4.0 or higher. Projects may use the CVSS score as published in a widely-used vulnerability database (such as the National Vulnerability Database) using the most-recent version of CVSS reported in that database. Projects may instead calculate the severity themselves using the latest version of CVSS at the time of the vulnerability disclosure, if the calculation inputs are publicly revealed once the vulnerability is publicly known. Note that criterion vulnerabilities_fixed_60_days requires that all such vulnerabilities be fixed within 60 days of being made public.

    No medium+ static analysis vulnerabilities exist. CI blocks merge until static analysis passes (Clippy, cargo-audit). Historical commits show same-day fixes for static analysis issues. CONTRIBUTING.md requires all Clippy checks pass before merge. Dependabot automates security updates.



    It is SUGGESTED that static source code analysis occur on every commit or at least daily. [static_analysis_often]

    Static analysis runs on every commit via GitHub Actions CI. Clippy, cargo-audit, cargo-deny, and CodeQL run on every push and PR. Pre-commit hooks run analysis locally before commit. Merging blocked until all static analysis passes. Additional daily scheduled security scans."


  • Dynamic code analysis


    It is SUGGESTED that at least one dynamic analysis tool be applied to any proposed major production release of the software before its release. [dynamic_analysis]
    A dynamic analysis tool examines the software by executing it with specific inputs. For example, the project MAY use a fuzzing tool (e.g., American Fuzzy Lop) or a web application scanner (e.g., OWASP ZAP or w3af). In some cases the OSS-Fuzz project may be willing to apply fuzz testing to your project. For purposes of this criterion the dynamic analysis tool needs to vary the inputs in some way to look for various kinds of problems or be an automated test suite with at least 80% branch coverage. The Wikipedia page on dynamic analysis and the OWASP page on fuzzing identify some dynamic analysis tools. The analysis tool(s) MAY be focused on looking for security vulnerabilities, but this is not required.

    Comprehensive dynamic analysis: cargo-fuzz with 7 fuzz targets (encryption, signatures, KEM, timing), proptest for property-based testing, daily scheduled fuzzing in CI. Runtime checks enabled: overflow-checks=true, debug-assertions=true in release builds.



    It is SUGGESTED that if the software produced by the project includes software written using a memory-unsafe language (e.g., C or C++), then at least one dynamic tool (e.g., a fuzzer or web application scanner) be routinely used in combination with a mechanism to detect memory safety problems such as buffer overwrites. If the project does not produce software written in a memory-unsafe language, choose "not applicable" (N/A). [dynamic_analysis_unsafe]
    Examples of mechanisms to detect memory safety problems include Address Sanitizer (ASAN) (available in GCC and LLVM), Memory Sanitizer, and valgrind. Other potentially-used tools include thread sanitizer and undefined behavior sanitizer. Widespread assertions would also work.

    LatticeArc is 100% memory-safe Rust with #![forbid(unsafe_code)]. No C/C++ code in the project. The only C dependency (aws-lc-rs) is FIPS 140-3 validated with
    its own memory safety testing (ASan/MSan/fuzzing) performed by AWS security team.



    It is SUGGESTED that the project use a configuration for at least some dynamic analysis (such as testing or fuzzing) which enables many assertions. In many cases these assertions should not be enabled in production builds. [dynamic_analysis_enable_assertions]
    This criterion does not suggest enabling assertions during production; that is entirely up to the project and its users to decide. This criterion's focus is instead to improve fault detection during dynamic analysis before deployment. Enabling assertions in production use is completely different from enabling assertions during dynamic analysis (such as testing). In some cases enabling assertions in production use is extremely unwise (especially in high-integrity components). There are many arguments against enabling assertions in production, e.g., libraries should not crash callers, their presence may cause rejection by app stores, and/or activating an assertion in production may expose private data such as private keys. Beware that in many Linux distributions NDEBUG is not defined, so C/C++ assert() will by default be enabled for production in those environments. It may be important to use a different assertion mechanism or defining NDEBUG for production in those environments.

    Assertions enabled for all dynamic analysis. Tests run in debug mode with debug_assertions=true. Release builds retain overflow-checks=true and
    debug-assertions=true for cryptographic validation. Fuzz targets run with full assertions. proptest verifies properties across random inputs with prop_assert macros.



    All medium and higher severity exploitable vulnerabilities discovered with dynamic code analysis MUST be fixed in a timely way after they are confirmed. [dynamic_analysis_fixed]
    If you are not running dynamic code analysis and thus have not found any vulnerabilities in this way, choose "not applicable" (N/A). A vulnerability is considered medium or higher severity if its Common Vulnerability Scoring System (CVSS) base qualitative score is medium or higher. In CVSS versions 2.0 through 3.1, this is equivalent to a CVSS score of 4.0 or higher. Projects may use the CVSS score as published in a widely-used vulnerability database (such as the National Vulnerability Database) using the most-recent version of CVSS reported in that database. Projects may instead calculate the severity themselves using the latest version of CVSS at the time of the vulnerability disclosure, if the calculation inputs are publicly revealed once the vulnerability is publicly known.

    No medium+ vulnerabilities discovered through dynamic analysis to date. 7 fuzz targets run daily in CI, with crash artifacts captured for analysis. Demonstrated rapid fix capability (3 releases in 2 days, same-day bug fixes). SECURITY.md documents response timelines: 7 days critical, 14 days high, 30 days medium.



This data is available under the Community Data License Agreement – Permissive, Version 2.0 (CDLA-Permissive-2.0). This means that a Data Recipient may share the Data, with or without modifications, so long as the Data Recipient makes available the text of this agreement with the shared Data. Please credit LatticeArc-Founder and the OpenSSF Best Practices badge contributors.

Project badge entry owned by: LatticeArc-Founder.
Entry created on 2026-01-30 15:38:28 UTC, last updated on 2026-01-31 05:50:07 UTC. Last achieved passing badge on 2026-01-31 05:44:05 UTC.