AI Declaration and Notes
LLMs and AI tools are a current fact of life. We feel it's important to be realistic and up front about how they affect this project.
Levels of AI Use
I created declare-ai.org to allow precise documentation of how AI is used in a project. It defines four levels of AI use:
- No involvement at all.
- Non-create assistance, such as the completions GitHub Copilot may make while you are typing.
- Creative assistance, where an AI generated a large amount of code that you then modified. An example would be using a tool like Cursor to write code for your review.
- Completely Produced by an AI, for example if you used Stable Diffusion to create your icon.
Code
The entirety of Brut's codebase could be grouped into three parts:
- The library code developers use. This is, roughly, the code in
lib
,brut-js/src
, andbrut-css/src
. - Tests of the library code.
- Tools used to manage the library code, such as what's in
bin/
,brut-js/bin
,brut-css/bin
, etc.
Library Code Should Be Written By a Person
We want the library code written by a person, i.e. it should have level 1 AI assistance: none. As of this writing—June 17, 2025—all library code was written by a person.
If some code is level 2—non-creative assistance—that is probably fine.
Tests Should Not Be Totally Written by AI
As of this writing—June 17, 2025—the library has no automated tests. The adrs.cloud
app serves as the test suite.
That said, as tests are developed, they must never be written entirely by an AI without human review (level 4). Our preference is that all tests comply with level 1 or level 2, however we are open to creative assistance for tests as might be provided by a tool like Cursor.
In any case, the author of a test should understand it and the maintainers must be able to understand and modify all tests without the use of AI tools.
Management Tools Should Not Be Totally Written by AI
Like tests, management tools must be comprehensible by a person and the creator of the code must understand it. That said, some of the tools were created with level 3 AI assistance, and we expect these tools may continue to be created this way.
Nevertheless, this code must be reviewed and understood by a person.
Documentation
We strive to have all documentation written by a real person. None of this documentation, or the API documentation, is produced by a machine, although spelling and grammar corrections will certainly have been suggested by automated tools.
We expect an LLM to be able to digest this documentation and source code and provide alternate analysis of how Brut works and how to use it. We hope such analysis is correct and useful, however that cannot be guaranteed, so this documentation is the second best source of truth, the source code being the best.
Logos
While the various logos look like someone typed "make me a logo in the style of the Washington, DC Metro", if you look closely, you will see the telltale sings that they were made by a person…a person who is not a professional designer.
As such, these are all Level 1.
AI Information about Brut Should Be Viewed with Skepticism
- Answers from an LLM about Brut are likely incorrect. LLMs will certainly not have been trained on information about Brut, since it is new.
- Code Completions of Brut Code are suspect. I have observed that e.g. GitHub CoPilot is capable for properly completing Brut code if there is enough context, but it is not capable of, say, creating a component from scratch. Review all Code Completions Carefully.