Deprecated: The each() function is deprecated. This message will be suppressed on further calls in /home/zhenxiangba/zhenxiangba.com/public_html/phproxy-improved-master/index.php on line 456
ReactiveAI (Reactive AI)
[go: Go Back, main page]

-->\n\nOur primary architecture - **Reactor** - is planned as the first _**awareness AGI model**_, that's modelling awareness as an _Infinite Chain-of-Thoughts_,\nconnected to _Short-Term and Long-Term Memory_ (_Attention-based Memory System_) and _Receptors/Effectors_ systems for real-time reactive processing.\nIt will be able to constantly and autonomously learn from interactions in _Continouos Live Learning_ process.\n\n> Reactor architecture details and mathematical model were analysed by 30 state-of-the-art LLM/Reasoning models that rated it's potential\n> to reach the AGI as ~4.35/5\n\n## Reactive Language Models (RxLM)\nWhile the **Reactor** is the main goal, it's extremely hard to achieve, as it's definitely the most advanced neural network ensemble ever.\n\nThat's why we designed simplified architectures, for incremental transformation from language/reasoning models to awareness model:\n- **Reactive Transformer** is introducing _Attention-based Memory System_ and adding _Short-Term Memory_ to Transformer language models\n- **Preactor** is adding _Long-Term Memory_ and ability to learn from interactions\n\n## RxLM vs LLM advantages\nProcessing single interactions in real-time by **Reactive Language Models** leads to **revolutional** improvements in inference speed/cost:\n- LLM inference costs are increasing quadratically with conversation length (accumulated for each next message), because of full dialog history processing\n- RxLM inference costs are linear, depending only on single interaction tokens (not accumulated) - each next interaction is `number of steps` times cheaper than for LLM\n- same for inference speed - LLM has to process full history, while RxLM only single message (only first interaction could be slower because of encoder/memory attention overhead)\n\n> In example, for a dialog with **DeepSeek R1**, that have overally ~90k tokens, I paid for about 1.5M tokens. With **RxLM** it will cost only that ~90k tokens, so it\n> will be about **15x cheaper**\n\n\n## RxNN Platform\n\n\n\n\n## Additional Research\n- **Sparse Query Attention (SQA)** - the most cost-effective GQA variant, even 2-3x faster for long sequences!\n- **Flex-SQA** - combination of Flex Attention and (symmetric) Sparse Query Attention, enabling 4-8x longer sliding windows\n- **Flex Memory Attention/Memory Cross-Attention** - connecting spatially sparse attention with memory layers to enable very long single interactions - smaller sliding window for input sequences attends to full memory, or the opposite\n- **Mixture-of-Experts for Grouped Attention** - MoE Router dynamically selects GQA/SQA groups, instead of static selection. Abandoned, because results were worse than for GQA/SQA","html":"\n\n

Reactive AI

\n

We are working on our own ideas of Reactive Neural Networks (RxNN) and Event-Driven AI, advancing from language models to AGI awareness models.

\n

Reactive Neural Networks and Event-Driven AI

\n

Reactive Neural Networks (RxNN) are memory-augmented neural networks with higher levels of recurrence (inter-sequence vs. intra-sequence in RNNs),\nfocused on processing single interactions with access to previous interactions via memory layers. We call this event-driven real-time processing\nto distinguish it from classical data-driven processing of the full conversation history in each interaction. This difference is crucial in case\nof AGI and awareness - the key feature of humans awareness, is that we remember what we were doing 10 mins ago, without recalling the whole-day history - we\nare working in real-time - just like event-driven Reactive Neural Networks.

\n

In Event-Driven AI models are processing the data in reaction to environment or internal events, and are emitting other response events as a result.\nProcessing of input and output events by the model is called the interaction. Event or an interaction could occur in any point in continous time. Models\nhave to be stateful and remember the data between the interactions.

\n

Strong Reactive Neural Networks like Reactor could emit and listen to its internal events, while the Weak Reactive Neural Networks are\nworking only on environment events.

\n

Reactor AGI

\n\n\n

Our primary architecture - Reactor - is planned as the first awareness AGI model, that's modelling awareness as an Infinite Chain-of-Thoughts,\nconnected to Short-Term and Long-Term Memory (Attention-based Memory System) and Receptors/Effectors systems for real-time reactive processing.\nIt will be able to constantly and autonomously learn from interactions in Continouos Live Learning process.

\n
\n

Reactor architecture details and mathematical model were analysed by 30 state-of-the-art LLM/Reasoning models that rated it's potential\nto reach the AGI as ~4.35/5

\n
\n

Reactive Language Models (RxLM)

\n

While the Reactor is the main goal, it's extremely hard to achieve, as it's definitely the most advanced neural network ensemble ever.

\n

That's why we designed simplified architectures, for incremental transformation from language/reasoning models to awareness model:

\n
    \n
  • Reactive Transformer is introducing Attention-based Memory System and adding Short-Term Memory to Transformer language models
  • \n
  • Preactor is adding Long-Term Memory and ability to learn from interactions
  • \n
\n

RxLM vs LLM advantages

\n

Processing single interactions in real-time by Reactive Language Models leads to revolutional improvements in inference speed/cost:

\n
    \n
  • LLM inference costs are increasing quadratically with conversation length (accumulated for each next message), because of full dialog history processing
  • \n
  • RxLM inference costs are linear, depending only on single interaction tokens (not accumulated) - each next interaction is number of steps times cheaper than for LLM
  • \n
  • same for inference speed - LLM has to process full history, while RxLM only single message (only first interaction could be slower because of encoder/memory attention overhead)
  • \n
\n
\n

In example, for a dialog with DeepSeek R1, that have overally ~90k tokens, I paid for about 1.5M tokens. With RxLM it will cost only that ~90k tokens, so it\nwill be about 15x cheaper

\n
\n

RxNN Platform

\n\n\n\n

Additional Research

\n
    \n
  • Sparse Query Attention (SQA) - the most cost-effective GQA variant, even 2-3x faster for long sequences!
  • \n
  • Flex-SQA - combination of Flex Attention and (symmetric) Sparse Query Attention, enabling 4-8x longer sliding windows
  • \n
  • Flex Memory Attention/Memory Cross-Attention - connecting spatially sparse attention with memory layers to enable very long single interactions - smaller sliding window for input sequences attends to full memory, or the opposite
  • \n
  • Mixture-of-Experts for Grouped Attention - MoE Router dynamically selects GQA/SQA groups, instead of static selection. Abandoned, because results were worse than for GQA/SQA
  • \n
\n","classNames":"hf-sanitized hf-sanitized-eussj193PyiFi_Da_1xnv"},"users":[{"_id":"675197c3ae96d7ba4b4a6c66","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/no-auth/I2GHgrv70cfT8C5EbK6Q5.png","isPro":true,"fullname":"Adam Filipek","user":"AdamF92","type":"user"},{"_id":"68e125bd10db6d257d004a96","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/no-auth/HrF2nHKJp6QonFFK4Gp8T.png","isPro":false,"fullname":"Weronika Bagniewska","user":"werbag","type":"user"},{"_id":"68e4cb221e68c13d99f271fc","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/no-auth/_MKd6YpjCvWCJO29UeExP.png","isPro":true,"fullname":"Wiktor Stanisławek","user":"Wstanislawek","type":"user"},{"_id":"6819e40283cb0783dde91b13","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/no-auth/U4K-v1YatVAQuCipLcAXT.png","isPro":false,"fullname":"Tomasz Filipek","user":"Tipifi","type":"user"}],"userCount":4,"collections":[{"slug":"ReactiveAI/rxt-beta-3b-a190m-reactive-transformer-mvp-69481125f8f6daad51fc5959","title":"RxT-Beta 3B A190M - Reactive Transformer MVP","description":"Reactive Transformer MVP model with 3B total params and 190M activated in decoder. Training in progress","gating":false,"lastUpdated":"2026-02-18T21:43:15.416Z","owner":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"items":[{"_id":"69690a78213899f567088155","position":0,"type":"model","author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":0,"gated":"auto","id":"ReactiveAI/RxT-Beta","availableInferenceProviders":[],"lastModified":"2026-02-16T13:26:45.000Z","likes":1,"pipeline_tag":"text-generation","private":false,"repoType":"model","isLikedByUser":false,"widgetOutputUrls":[]},{"_id":"694811ca2874630ec1e041a9","position":1,"type":"model","author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":116,"gated":"auto","id":"ReactiveAI/RxT-Beta-Decoder-Base","availableInferenceProviders":[],"lastModified":"2026-02-18T21:40:01.000Z","likes":0,"pipeline_tag":"text-generation","private":false,"repoType":"model","isLikedByUser":false,"widgetOutputUrls":[],"numParameters":2886109200},{"_id":"694811d105a6867444bc0837","position":2,"type":"model","author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":115,"gated":"auto","id":"ReactiveAI/RxT-Beta-Encoder-Base","availableInferenceProviders":[],"lastModified":"2026-02-18T21:40:35.000Z","likes":1,"pipeline_tag":"fill-mask","private":false,"repoType":"model","isLikedByUser":false,"widgetOutputUrls":[],"numParameters":97090064},{"_id":"694811d901886522f021eda7","position":3,"type":"model","author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":106,"gated":"auto","id":"ReactiveAI/RxT-Beta-MLM-Base","availableInferenceProviders":[],"lastModified":"2026-02-18T21:39:21.000Z","likes":0,"pipeline_tag":"fill-mask","private":false,"repoType":"model","isLikedByUser":false,"widgetOutputUrls":[]}],"position":0,"theme":"orange","private":false,"shareUrl":"https://hf.co/collections/ReactiveAI/rxt-beta-3b-a190m-reactive-transformer-mvp","upvotes":0,"isUpvotedByUser":false},{"slug":"ReactiveAI/reactive-transformer-poc-rxt-beta-micro-models-6908e4afd6f3c4e54df529d9","title":"Reactive Transformer PoC - RxT-Beta-Micro models","description":"Second generation of Reactive Transformer (RxT) models, trained on limited english-only datasets","gating":false,"lastUpdated":"2025-12-21T15:22:43.448Z","owner":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"items":[{"_id":"6908e567fee25ab31af70c16","position":0,"type":"model","author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":0,"gated":"auto","id":"ReactiveAI/RxT-Beta-Micro-Supervised-AI","availableInferenceProviders":[],"lastModified":"2025-11-04T15:56:56.000Z","likes":0,"pipeline_tag":"text-generation","private":false,"repoType":"model","isLikedByUser":false,"widgetOutputUrls":[],"numParameters":282555928},{"_id":"6908e55e916c6f569f2035d5","position":1,"type":"model","author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":37,"gated":"auto","id":"ReactiveAI/RxT-Beta-Micro-Supervised","availableInferenceProviders":[],"lastModified":"2025-11-19T11:11:21.000Z","likes":0,"pipeline_tag":"text-generation","private":false,"repoType":"model","isLikedByUser":false,"widgetOutputUrls":[],"numParameters":300516632},{"_id":"69094cec80107c8ae3e9c315","position":2,"type":"paper","id":"2510.03561","title":"Reactive Transformer (RxT) -- Stateful Real-Time Processing for\n Event-Driven Reactive Language Models","thumbnailUrl":"https://cdn-thumbnails.huggingface.co/social-thumbnails/papers/2510.03561.png","upvotes":25,"publishedAt":"2025-10-03T23:18:07.000Z","isUpvotedByUser":false}],"position":1,"theme":"indigo","private":false,"shareUrl":"https://hf.co/collections/ReactiveAI/reactive-transformer-poc-rxt-beta-micro-models","upvotes":0,"isUpvotedByUser":false},{"slug":"ReactiveAI/rxt-beta-training-datasets-68ea911ed219d9ae11ba9f6d","title":"RxT-Beta Training Datasets","description":"Datasets used to train RxT-Beta models - first generation of experimental Reactive Transformer (RxT) models trained on real-world data (English only)","gating":false,"lastUpdated":"2025-12-21T15:23:56.655Z","owner":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"items":[{"_id":"6903746b6c3935cdb568d477","position":0,"type":"dataset","author":"ReactiveAI","downloads":115,"gated":false,"id":"ReactiveAI/AI-Knowledge-Chat-SMAT","lastModified":"2025-10-29T00:15:15.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":22884,"libraries":["datasets","pandas","mlcroissant","polars"],"formats":["parquet"],"modalities":["text"]},"private":false,"repoType":"dataset","likes":2,"isLikedByUser":false,"isBenchmark":false},{"_id":"69037483b97fce183dab3ffe","position":1,"type":"dataset","author":"ReactiveAI","downloads":32,"gated":false,"id":"ReactiveAI/Real-Chat-No-System-SMAT","lastModified":"2025-10-25T15:01:04.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":28210,"libraries":["datasets","pandas","mlcroissant","polars"],"formats":["parquet"],"modalities":["text"]},"private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false},{"_id":"69037477d756c51b1f758df7","position":2,"type":"dataset","author":"ReactiveAI","downloads":74,"gated":false,"id":"ReactiveAI/Real-Chat-SMAT","lastModified":"2025-10-12T16:14:56.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":84157,"libraries":["datasets","dask","mlcroissant","polars"],"formats":["parquet"],"modalities":["text"]},"private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false},{"_id":"6903749ab25c5202aa087999","position":3,"type":"dataset","author":"ReactiveAI","downloads":100,"gated":false,"id":"ReactiveAI/Wild-Chat-MRL","lastModified":"2025-10-22T14:33:16.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":33202,"libraries":["datasets","pandas","mlcroissant","polars"],"formats":["parquet"],"modalities":["text"]},"private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false}],"position":2,"theme":"orange","private":false,"shareUrl":"https://hf.co/collections/ReactiveAI/rxt-beta-training-datasets","upvotes":0,"isUpvotedByUser":false},{"slug":"ReactiveAI/reactive-transformer-poc-rxt-alpha-supervised-models-68e4004a4a59366e01a7b86f","title":"Reactive Transformer PoC - RxT-Alpha Supervised Models","description":"Experimental stateful real-time Reactive Transformer (RxT) models after supervised training stages","gating":false,"lastUpdated":"2025-10-08T13:08:52.802Z","owner":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"items":[{"_id":"68e4708cb3c0bac9940220f5","position":0,"type":"paper","id":"2510.03561","title":"Reactive Transformer (RxT) -- Stateful Real-Time Processing for\n Event-Driven Reactive Language Models","thumbnailUrl":"https://cdn-thumbnails.huggingface.co/social-thumbnails/papers/2510.03561.png","upvotes":25,"publishedAt":"2025-10-03T23:18:07.000Z","isUpvotedByUser":false},{"_id":"68e6625c9b2c9aec120b908f","position":1,"type":"model","author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":0,"gated":"auto","id":"ReactiveAI/RxT-Alpha-Supervised","availableInferenceProviders":[],"lastModified":"2025-10-08T13:08:10.000Z","likes":0,"pipeline_tag":"text-generation","private":false,"repoType":"model","isLikedByUser":false,"widgetOutputUrls":[],"numParameters":163076408},{"_id":"68e462ae4bb9f975ac572419","position":2,"type":"model","author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":0,"gated":"auto","id":"ReactiveAI/RxT-Alpha-Mini-Supervised","availableInferenceProviders":[],"lastModified":"2025-10-07T01:45:56.000Z","likes":0,"pipeline_tag":"text-generation","private":false,"repoType":"model","isLikedByUser":false,"widgetOutputUrls":[],"numParameters":103092776},{"_id":"68e4008523e838a4741d9f5d","position":3,"type":"model","author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":0,"gated":"auto","id":"ReactiveAI/RxT-Alpha-Micro-Supervised","availableInferenceProviders":[],"lastModified":"2025-10-07T01:46:05.000Z","likes":0,"pipeline_tag":"text-generation","private":false,"repoType":"model","isLikedByUser":false,"widgetOutputUrls":[],"numParameters":28805092}],"position":3,"theme":"indigo","private":false,"shareUrl":"https://hf.co/collections/ReactiveAI/reactive-transformer-poc-rxt-alpha-supervised-models","upvotes":2,"isUpvotedByUser":false},{"slug":"ReactiveAI/sparse-query-attention-sqa-research-6813d6e8315e6ce20d22a37b","title":"Sparse Query Attention (SQA) Research","description":"Experimental models with Sparse Query Attention layers. Reducing training time/cost by ~3-10% compared to GQA & MQA, with the same level performance","gating":true,"lastUpdated":"2025-10-03T16:32:56.686Z","owner":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"items":[{"_id":"68dffa26d47e38372548ae7d","position":0,"type":"paper","id":"2510.01817","title":"Sparse Query Attention (SQA): A Computationally Efficient Attention\n Mechanism with Query Heads Reduction","thumbnailUrl":"https://cdn-thumbnails.huggingface.co/social-thumbnails/papers/2510.01817.png","upvotes":16,"publishedAt":"2025-10-02T09:01:38.000Z","isUpvotedByUser":false},{"_id":"681cbea3446ded3bf98c19d1","position":1,"type":"model","author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":0,"gated":"auto","id":"ReactiveAI/sSQAT-mm","availableInferenceProviders":[],"lastModified":"2025-10-03T16:34:47.000Z","likes":0,"pipeline_tag":"text-generation","private":false,"repoType":"model","isLikedByUser":false,"widgetOutputUrls":[],"numParameters":8619152},{"_id":"681cbe9ced46c7219e41c3ac","position":2,"type":"model","author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":0,"gated":"auto","id":"ReactiveAI/SQAT-mm","availableInferenceProviders":[],"lastModified":"2025-10-03T16:35:26.000Z","likes":0,"pipeline_tag":"text-generation","private":false,"repoType":"model","isLikedByUser":false,"widgetOutputUrls":[],"numParameters":8570000},{"_id":"681cbea9063e1a8e696f3706","position":3,"type":"model","author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":0,"gated":"auto","id":"ReactiveAI/xSQAT-mm","availableInferenceProviders":[],"lastModified":"2025-10-03T16:35:40.000Z","likes":0,"pipeline_tag":"text-generation","private":false,"repoType":"model","isLikedByUser":false,"widgetOutputUrls":[],"numParameters":8520848}],"position":6,"theme":"orange","private":false,"shareUrl":"https://hf.co/collections/ReactiveAI/sparse-query-attention-sqa-research","upvotes":1,"isUpvotedByUser":false},{"slug":"ReactiveAI/interaction-sft-datasets-68288f2c0cc5c7716a073798","title":"Interaction SFT Datasets","description":"Datasets used for Interaction Supervised Fine-Tuning (SFT) of reactive models, made for real-time processing of single sequence (interaction)","gating":false,"lastUpdated":"2025-10-11T17:16:29.377Z","owner":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"items":[{"_id":"68288feb542dc279d6ebfca4","position":0,"type":"dataset","author":"ReactiveAI","downloads":144,"gated":false,"id":"ReactiveAI/TinyStories-mini-Interaction-SFT","lastModified":"2025-05-17T13:37:03.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":24320,"libraries":["datasets","pandas","mlcroissant","polars"],"formats":["parquet"],"modalities":["text"]},"private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false},{"_id":"68288ffeb1a447428bdb94c8","position":1,"type":"dataset","author":"ReactiveAI","downloads":17,"gated":false,"id":"ReactiveAI/TinyStories-Interaction-SFT","lastModified":"2025-05-31T22:47:06.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":118452,"libraries":["datasets","pandas","mlcroissant","polars"],"formats":["parquet"],"modalities":["text"]},"private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false},{"_id":"68289008717838b1591e9574","position":2,"type":"dataset","author":"ReactiveAI","downloads":419,"gated":false,"id":"ReactiveAI/smol-smoltalk-Interaction-SFT","lastModified":"2025-10-08T22:38:44.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":1085969,"libraries":["datasets","dask","mlcroissant","polars"],"formats":["parquet"],"modalities":["text"]},"private":false,"repoType":"dataset","likes":2,"isLikedByUser":false,"isBenchmark":false},{"_id":"68289015ea6eb78a8eda1f48","position":3,"type":"dataset","author":"ReactiveAI","downloads":12,"gated":false,"id":"ReactiveAI/glaive-reasoning-Interaction-SFT","lastModified":"2025-10-07T00:22:42.000Z","datasetsServerInfo":{"viewer":"preview","numRows":0,"libraries":[],"formats":[],"modalities":[]},"private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false}],"position":9,"theme":"green","private":false,"shareUrl":"https://hf.co/collections/ReactiveAI/interaction-sft-datasets","upvotes":0,"isUpvotedByUser":false}],"datasets":[{"author":"ReactiveAI","downloads":40,"gated":false,"id":"ReactiveAI/coqa-retrieval","lastModified":"2026-02-21T14:08:38.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":6585,"libraries":["datasets","pandas","polars","mlcroissant"],"formats":["parquet"],"modalities":["text"]},"private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false},{"author":"ReactiveAI","downloads":1573,"gated":false,"id":"ReactiveAI/beta-reasoning","lastModified":"2026-02-21T14:05:29.000Z","private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false},{"author":"ReactiveAI","downloads":45,"gated":false,"id":"ReactiveAI/passkey-retrieval","lastModified":"2026-02-20T22:29:35.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":1571,"libraries":["datasets","pandas","polars","mlcroissant"],"formats":["parquet"],"modalities":["text"]},"private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false},{"author":"ReactiveAI","downloads":184,"gated":false,"id":"ReactiveAI/Beta-Hybrid-Interaction-SFT","lastModified":"2026-02-19T23:14:11.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":55663989,"libraries":["datasets","dask","polars","mlcroissant"],"formats":["parquet"],"modalities":["text"]},"private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false},{"author":"ReactiveAI","downloads":13,"gated":false,"id":"ReactiveAI/codeparrot-apps-reupload","lastModified":"2026-01-30T12:30:11.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":10000,"libraries":["datasets","dask","polars","mlcroissant"],"formats":["parquet"],"modalities":["text"]},"private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false},{"author":"ReactiveAI","downloads":24,"gated":false,"id":"ReactiveAI/BAAI-TACO-reupload","lastModified":"2026-01-30T12:29:26.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":26443,"libraries":["datasets","dask","polars","mlcroissant"],"formats":["parquet"],"modalities":["text"]},"private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false},{"author":"ReactiveAI","downloads":14211,"gated":false,"id":"ReactiveAI/Beta-Pre-Train-Corpus","lastModified":"2026-01-27T20:42:31.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":333034301,"libraries":["datasets","dask","polars","mlcroissant"],"formats":["parquet","optimized-parquet"],"modalities":["text"]},"private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false},{"author":"ReactiveAI","downloads":685,"gated":false,"id":"ReactiveAI/algebraic-stack-fixed","lastModified":"2025-12-17T18:06:43.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":3391835,"libraries":["datasets","dask","mlcroissant","polars"],"formats":["parquet"],"modalities":["text"]},"private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false},{"author":"ReactiveAI","downloads":2012,"gated":false,"id":"ReactiveAI/Beta-Code","lastModified":"2025-12-17T16:53:38.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":6146498,"libraries":["datasets","dask","mlcroissant","polars"],"formats":["parquet"],"modalities":["text"]},"private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false},{"author":"ReactiveAI","downloads":235,"gated":false,"id":"ReactiveAI/finepdfs-edu-beta","lastModified":"2025-11-15T23:55:43.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":13664942,"libraries":["datasets","dask","mlcroissant","polars"],"formats":["parquet"],"modalities":["tabular","text"]},"private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false}],"models":[{"author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":173,"gated":false,"id":"ReactiveAI/RxT-Beta-MLM-iSFT","availableInferenceProviders":[],"lastModified":"2026-02-20T10:10:44.000Z","likes":0,"private":false,"repoType":"model","isLikedByUser":false},{"author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":192,"gated":false,"id":"ReactiveAI/RxT-Beta-Encoder-iSFT","availableInferenceProviders":[],"lastModified":"2026-02-20T10:10:39.000Z","likes":0,"private":false,"repoType":"model","isLikedByUser":false,"numParameters":97090064},{"author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":187,"gated":false,"id":"ReactiveAI/RxT-Beta-Decoder-iSFT","availableInferenceProviders":[],"lastModified":"2026-02-20T10:10:30.000Z","likes":0,"private":false,"repoType":"model","isLikedByUser":false,"numParameters":2930149392},{"author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":251,"gated":false,"id":"ReactiveAI/RxT-Beta-Mini-Encoder-Base","availableInferenceProviders":[],"lastModified":"2026-02-19T17:30:39.000Z","likes":0,"private":false,"repoType":"model","isLikedByUser":false,"numParameters":57399180},{"author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":274,"gated":false,"id":"ReactiveAI/RxT-Beta-Mini-Decoder-Base","availableInferenceProviders":[],"lastModified":"2026-02-19T17:30:30.000Z","likes":0,"private":false,"repoType":"model","isLikedByUser":false,"numParameters":1161852172},{"author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":248,"gated":false,"id":"ReactiveAI/RxT-Beta-Mini-MLM-Base","availableInferenceProviders":[],"lastModified":"2026-02-19T17:29:36.000Z","likes":0,"private":false,"repoType":"model","isLikedByUser":false},{"author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":115,"gated":"auto","id":"ReactiveAI/RxT-Beta-Encoder-Base","availableInferenceProviders":[],"lastModified":"2026-02-18T21:40:35.000Z","likes":1,"pipeline_tag":"fill-mask","private":false,"repoType":"model","isLikedByUser":false,"widgetOutputUrls":[],"numParameters":97090064},{"author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":116,"gated":"auto","id":"ReactiveAI/RxT-Beta-Decoder-Base","availableInferenceProviders":[],"lastModified":"2026-02-18T21:40:01.000Z","likes":0,"pipeline_tag":"text-generation","private":false,"repoType":"model","isLikedByUser":false,"widgetOutputUrls":[],"numParameters":2886109200},{"author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":106,"gated":"auto","id":"ReactiveAI/RxT-Beta-MLM-Base","availableInferenceProviders":[],"lastModified":"2026-02-18T21:39:21.000Z","likes":0,"pipeline_tag":"fill-mask","private":false,"repoType":"model","isLikedByUser":false,"widgetOutputUrls":[]},{"author":"ReactiveAI","authorData":{"_id":"675776b060e4100500aeb4c8","avatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","fullname":"Reactive AI","name":"ReactiveAI","type":"org","isHf":false,"isHfAdmin":false,"isMod":false,"followerCount":12,"isUserFollowing":false},"downloads":0,"gated":"auto","id":"ReactiveAI/RxT-Beta","availableInferenceProviders":[],"lastModified":"2026-02-16T13:26:45.000Z","likes":1,"pipeline_tag":"text-generation","private":false,"repoType":"model","isLikedByUser":false,"widgetOutputUrls":[]}],"paperPreviews":[{"_id":"2510.05485","title":"TensorBLEU: Vectorized GPU-based BLEU Score Implementation for\n Per-Sentence In-Training Evaluation","id":"2510.05485","thumbnailUrl":"https://cdn-thumbnails.huggingface.co/social-thumbnails/papers/2510.05485.png"},{"_id":"2510.03561","title":"Reactive Transformer (RxT) -- Stateful Real-Time Processing for\n Event-Driven Reactive Language Models","id":"2510.03561","thumbnailUrl":"https://cdn-thumbnails.huggingface.co/social-thumbnails/papers/2510.03561.png"}],"spaces":[],"buckets":[],"numBuckets":0,"numDatasets":34,"numModels":28,"numSpaces":1,"lastOrgActivities":[{"time":"2026-02-21T00:02:44.623Z","user":"AdamF92","userAvatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/no-auth/I2GHgrv70cfT8C5EbK6Q5.png","orgAvatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","type":"update","repoData":{"author":"ReactiveAI","downloads":40,"gated":false,"id":"ReactiveAI/coqa-retrieval","lastModified":"2026-02-21T14:08:38.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":6585,"libraries":["datasets","pandas","polars","mlcroissant"],"formats":["parquet"],"modalities":["text"]},"private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false},"repoId":"ReactiveAI/coqa-retrieval","repoType":"dataset","org":"ReactiveAI"},{"time":"2026-02-21T00:01:06.487Z","user":"AdamF92","userAvatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/no-auth/I2GHgrv70cfT8C5EbK6Q5.png","orgAvatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","type":"update","repoData":{"author":"ReactiveAI","downloads":1573,"gated":false,"id":"ReactiveAI/beta-reasoning","lastModified":"2026-02-21T14:05:29.000Z","private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false},"repoId":"ReactiveAI/beta-reasoning","repoType":"dataset","org":"ReactiveAI"},{"time":"2026-02-20T21:09:16.086Z","user":"AdamF92","userAvatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/no-auth/I2GHgrv70cfT8C5EbK6Q5.png","orgAvatarUrl":"https://cdn-avatars.huggingface.co/v1/production/uploads/675197c3ae96d7ba4b4a6c66/AJDkLuavcYfENIRDzxjqR.png","type":"publish","repoData":{"author":"ReactiveAI","downloads":40,"gated":false,"id":"ReactiveAI/coqa-retrieval","lastModified":"2026-02-21T14:08:38.000Z","datasetsServerInfo":{"viewer":"viewer","numRows":6585,"libraries":["datasets","pandas","polars","mlcroissant"],"formats":["parquet"],"modalities":["text"]},"private":false,"repoType":"dataset","likes":0,"isLikedByUser":false,"isBenchmark":false},"repoId":"ReactiveAI/coqa-retrieval","repoType":"dataset","org":"ReactiveAI"}],"acceptLanguages":["*"],"canReadRepos":false,"canReadSpaces":false,"blogPosts":[],"currentRepoPage":0,"filters":{},"paperView":false}">

AI & ML interests

AGI, ASI, Reactive Awareness Models, Real-Time Reactive Language Models, Memory Systems, Reactive Neural Networks & Event-Driven AI

Recent Activity

AdamF92  updated a dataset about 14 hours ago
ReactiveAI/coqa-retrieval
AdamF92  updated a dataset about 14 hours ago
ReactiveAI/beta-reasoning
AdamF92  published a dataset about 17 hours ago
ReactiveAI/coqa-retrieval
View all activity

Reactive AI

We are working on our own ideas of Reactive Neural Networks (RxNN) and Event-Driven AI, advancing from language models to AGI awareness models.

Reactive Neural Networks and Event-Driven AI

Reactive Neural Networks (RxNN) are memory-augmented neural networks with higher levels of recurrence (inter-sequence vs. intra-sequence in RNNs), focused on processing single interactions with access to previous interactions via memory layers. We call this event-driven real-time processing to distinguish it from classical data-driven processing of the full conversation history in each interaction. This difference is crucial in case of AGI and awareness - the key feature of humans awareness, is that we remember what we were doing 10 mins ago, without recalling the whole-day history - we are working in real-time - just like event-driven Reactive Neural Networks.

In Event-Driven AI models are processing the data in reaction to environment or internal events, and are emitting other response events as a result. Processing of input and output events by the model is called the interaction. Event or an interaction could occur in any point in continous time. Models have to be stateful and remember the data between the interactions.

Strong Reactive Neural Networks like Reactor could emit and listen to its internal events, while the Weak Reactive Neural Networks are working only on environment events.

Reactor AGI

Our primary architecture - Reactor - is planned as the first awareness AGI model, that's modelling awareness as an Infinite Chain-of-Thoughts, connected to Short-Term and Long-Term Memory (Attention-based Memory System) and Receptors/Effectors systems for real-time reactive processing. It will be able to constantly and autonomously learn from interactions in Continouos Live Learning process.

Reactor architecture details and mathematical model were analysed by 30 state-of-the-art LLM/Reasoning models that rated it's potential to reach the AGI as ~4.35/5

Reactive Language Models (RxLM)

While the Reactor is the main goal, it's extremely hard to achieve, as it's definitely the most advanced neural network ensemble ever.

That's why we designed simplified architectures, for incremental transformation from language/reasoning models to awareness model:

  • Reactive Transformer is introducing Attention-based Memory System and adding Short-Term Memory to Transformer language models
  • Preactor is adding Long-Term Memory and ability to learn from interactions

RxLM vs LLM advantages

Processing single interactions in real-time by Reactive Language Models leads to revolutional improvements in inference speed/cost:

  • LLM inference costs are increasing quadratically with conversation length (accumulated for each next message), because of full dialog history processing
  • RxLM inference costs are linear, depending only on single interaction tokens (not accumulated) - each next interaction is number of steps times cheaper than for LLM
  • same for inference speed - LLM has to process full history, while RxLM only single message (only first interaction could be slower because of encoder/memory attention overhead)

In example, for a dialog with DeepSeek R1, that have overally ~90k tokens, I paid for about 1.5M tokens. With RxLM it will cost only that ~90k tokens, so it will be about 15x cheaper

RxNN Platform

Additional Research

  • Sparse Query Attention (SQA) - the most cost-effective GQA variant, even 2-3x faster for long sequences!
  • Flex-SQA - combination of Flex Attention and (symmetric) Sparse Query Attention, enabling 4-8x longer sliding windows
  • Flex Memory Attention/Memory Cross-Attention - connecting spatially sparse attention with memory layers to enable very long single interactions - smaller sliding window for input sequences attends to full memory, or the opposite
  • Mixture-of-Experts for Grouped Attention - MoE Router dynamically selects GQA/SQA groups, instead of static selection. Abandoned, because results were worse than for GQA/SQA