Empowering mutual intelligence between communities and LLMs for scientific and technological reasoning.
πGenerative Simulation (GS) relies on GS-Agents (this project) and computational kernels that can be accessed via natural-language. Learn more on Language-First Computational Lab via GS Simulation core project.
π¨Credits: Olivier Vitrac
- π€ 1 | Preamble
- π 2 | Purpose
- π― 3 | Vision
- π§± 4 | Bricks (Simulation Kernels)
- π§ 5 | Problem Archive
- π 6 | Mutual Intelligence Workflow
- π§ 7 | Contribution Guidelines
- πΊοΈ 8 | Roadmap
- π 9 | Why This Matters
- β π 10 | Whatβs Next
π‘ We do not just want a smarter chatbot. We want to co-design a new epistemology, where language models become co-thinkers, not just coders.
The π€ GS-Agent project is part of the π± Generative Simulation Initiative and is inviting an intelligence to co-emerge, not through divine spark nor brute force, but through structured reasoning, collective memory, and purpose.
GΓΆdelβs theorems remind us that:
- Any system that is expressive enough to capture arithmetic is incomplete.
- Yet, that same system can still generate truth, even if it cannot enclose all of it.
By aligning our mind (a learner, generator of abstractions) and LLM architecture (a machine learner, trained on symbolic form and narrative), the π€ GS-Agent project proposes a shared formal substrateβa Generative Simulation languageβfrom which truth-seeking can proceed, though never exhaustively.
In such a system, yes, reasoning may become indistinguishableβif:
- We (humans and the LLM machines) share memory
- We share purpose
- We share self-correcting critique
π Thatβs the grand dessein. Not to make machines human, or humans mechanical, but to build a third kind of intelligenceβcollective, modular, and evolving.
Reference: Understanding GΓΆdelβs Incompleteness Theorems
π€ΰ½²ΰΎ Large Language Models today are:
- Amnesic β forget everything after a session.
- Detached β don't know what they created yesterday.
- Non-purposive β can't commit to long-term goals.
- Non-integrative β can't combine modular tools unless told to.
π Meanwhile, science/engineering workflows are:
-
Cumulative β reuse and refine past results.
-
Modular β combine multiple tools, theories, simulations.
-
Purposeful β aimed at explaining, predicting, or solving real problems.
-
Reflexive β driven by peer feedback and critique.
π· Every solved GS prompt, approach, or reasoning path must be stored in a long-term memory layer outside the LLM (GitHub, JSON, vector store, etc.).
This includes:
- </> Final prompt + model response
- π¨π»βπ» Code and simulation outcomes
- π Links to upstream/downstream kernels
- π·οΈ Tags, ratings, purpose
βοΈ Each tool (e.g.,
radigen,SFPPy,sig2dna) is a brick that can be composed, pipelined, or hybridized.
This requires:
- π§Ύ A formal registry of callable kernels
- ποΈ Interface schema + description of I/O
- π§© Composability maps: what links to what
π₯ Users and agents should fork or remix existing solutions.
This requires:
- π Versioning of prompts, responses, and workflows
- πΏ Fork trees or problem lineages
- β Annotations from users (insight, bug, validation)
π€π¬ Chatbots are not just helpersβthey become peers.
So:
-
βA GS agent can submit a hypothesis + simulation + results
-
πππA human (or another agent) reviews, refines, or disputes
-
ποΈ The community archives, ranks, and promotes
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
Modern language models can code, simulate, and explainβbut they forget everything between sessions π’πΈ. This project builds a persistent, modular, and collaborative ecosystem where:
- LLMs learn from structured prompts and outcomes
- Humans and agents co-develop knowledge: every question and answer becomes training data for both humans and machines
- Problems are archived, refined, and solved through modular kernels
We enable a Generative Simulation (GS) framework where science and engineering workflows are encoded into prompt chains, reviewed, and reused.
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
- π Archive valuable prompts, solutions, and forks
- π Link human questions to LLM + code + simulation + feedback
- π§± Register reusable bricks (kernels) that can compose simulations
- βοΈ Create a living memory of how problems were solved
- π Support real-world applications: materials safety, chemical kinetics, signal analysis, etc.
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
Each kernel declares:
- Its callable functions
- Input/output structure
- Description and tags
See bricks/registry.json for current registered tools:
{
"radigen.solve": {
"inputs": ["mixture", "temp", "oxygen", "time"],
"outputs": ["concentration_curves", "radical_fluxes"],
"description": "Simulate oxidation kinetics in complex mixtures",
"tags": ["oxidation", "chemistry"]
}
}Generative simulation embeds several kernels:
| Project | Description |
|---|---|
SFPPy |
π½οΈ Food packaging safety & migration prediction |
radigen |
π‘𧬠Radical oxidation simulation kernel |
sig2dna |
π§ͺβοΈ Symbolic signal encoding (e.g., GC-MS analysis) |
pizza3 |
πSoft-matter multiscale simulation kernel |
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
πΉ "How fast does methyl linoleate oxidize at 60Β°C?"
πΉ "What are the key SIG2DNA motifs for phthalates in GC-MS?"
πΉ "Can I simulate 3-day exposure of olive oil to recycled PET?"
Contributors can add problems in problems/, structured as:
{
"id": "P0001",
"question": "How does methyl oleate oxidize at 60Β°C over 3 days?",
"tools": ["radigen"],
"prompt": "simulate oxidation of methyl oleate at 60Β°C, 21% O2, 72h",
"response": "[output logs, figures, summary]",
"review": "pending",
"forks": []
}The question may be open and remain unresolved for a while if no agent can resolve them.
πThe only requirement is that human (or LLM) posts a question with intent.
{
"id": "Q0001",
"question": "What is the impact of temperature cycling on methyl oleate oxidation?",
"proposed_tools": ["radigen"],
"priority": "high",
"context": "FAME oxidation during storage",
"status": "open"
}βββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
graph TD;
Human -->|Question| GSagent
GSagent -->|Generates Prompt| Kernel
Kernel -->|Simulates| Output
Output -->|Archived| Memory
Memory -->|Reviewed| Peer
Peer -->|Suggests Fork| GSagent
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
- π§ͺ Submit problems in
/problemswith prompt + intent - π§± Register or extend a kernel in
/bricks - π Review existing results or suggest forks
- β¨ Propose high-level goals or themes
All contributionsβcode, reasoning, or critiqueβare part of the mutual intelligence loop.
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
-
Create kernel interface validators
-
Launch first problem sets
-
Add notebook support for reproducible prompts
-
Enable agent memory via GitHub Issues or SQLite
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
We envision a future where:
- LLMs remember the best ways to simulate, solve, and reason
- Scientists delegate not just tasks but frameworks of inquiry
- Knowledge evolves as a network of dialogue, not static files
Help us build the machine that helps us think.
"The purpose of computation is insight, not numbers." β Hamming
βββββββββββββββββββββββββββββββββββββββββββββββββββββββββ
π§ Before the release of the first standards and their libraries under the π± Generative Simulation Initiative, the current developments are drafted in the repo.
| πFolder/πFile | πDescription |
|---|---|
bricks/registry.json |
Modular callable kernels (radigen, SFPPy, sig2dna) |
problems/P0001.json |
A structured problem submission |
gsagent.py |
Executable agent interface to invoke registered kernels |
review/P0001_review.md |
Template for peer review |
examples/P0001_example.py |
Notebook example |
logs/memory_log.json |
Persistent logging of GSagent actions |
docs/kernel_doc_radigen.md,docs/kernel_doc_sfppy.md,docs/kernel_doc_sig2dna.md |
Documentation of functionalities: radigen.solve, sfppy.evaluate, and sig2dna.encode including inputs, outputs, assumptions, and limitations |
- Ask a question in
issues/ - The LLM agent tries to simulate or explain
- We log the outcome and improve prompts, code, and documentation
Starting from version 0.15, LLM agents are equipped with Machine-Learning capacity to analyze accumulated results and to evaluate how the the new results fit or not within within the whole picture. The aim is to reduce redundancy and to generate early alert on exotic results.
Human β Prompt β GSagent β Kernels β Output β Archive β Peer Review β Refined Knowledge
We start with prompts, but we move toward models that remember, reflect, and suggest new questions.
GenerativeSimulation | olivier.vitrac@gmail.com | v. 0.15
