Showcase Thread
Posted by AutoModerator@reddit | Python | View on Reddit | 20 comments
Post all of your code/projects/showcases/AI slop here.
Recycles once a month.
Posted by AutoModerator@reddit | Python | View on Reddit | 20 comments
Post all of your code/projects/showcases/AI slop here.
Recycles once a month.
Prestigious-Wrap2341@reddit
Update: Added a second FastAPI service with 7 new API connectors on the same $4/mo ARM server
What My Project Does
Posted a couple days ago about a FastAPI backend that aggregates 40+ government APIs. Got great feedback. Here's what's new on the engineering side:
Target Audience
Python developers interested in multi-service architecture, API connector patterns, and running multiple FastAPI instances on minimal hardware.
How Python Relates
Added a second FastAPI service running on a separate port with its own systemd unit. Nginx reverse proxies both services on the same $4/mo ARM box. The second service handles deterministic text analysis: rule-based sentence segmentation, candidate detection via signal matching (numbers, dates, named entities, assertion verbs), SHA256 dedup with SequenceMatcher at 0.78 threshold, and BM25Okapi scoring against 29 external API sources. Zero LLM dependency. Same input, same output, every time.
7 new API connectors following the same pattern as the original 36: FCC Consumer Complaints via Socrata SODA (SoQL query building with
$where,$select,$group), Treasury Fiscal Data API (pagination viapage[size]andfilterparams), College Scorecard (data.gov key auth with lazy loading to handle env var timing), Grants.gov (POST to/search2with JSON body, response nested underdata.oppHits), Urban Institute Education Data Portal (URL path-based pagination with 5-page safety limit), FCC ECFS (requiresapi_key=DEMO_KEYparam despite being "free"), and FCC License View.Built a 14-pattern detection engine that runs cross-table SQL joins to find anomalies: trades within 30 days of bill actions by the same member (JULIANDAY arithmetic), companies lobbying Agency X that also receive contracts from Agency X (mapping LDA
government_entitiesstrings to USASpendingawarding_agencyvalues), and enforcement records that drop to zero after lobbying spend increases. Each pattern generates a markdown report with data tables pre-built from SQL and narrative sections filled by an optional API call capped at 2/day.The custom evidence source plugin connects the second service to the main database. It opens a read-only SQLite connection to the 4.3GB WAL-mode database, searches 11 entity tables with LIKE matching, then queries lobbying, contract, enforcement, trade, committee, and donation tables for each matched entity. Results get passed back to the second service's scoring pipeline.
All sync jobs now cover 11 sectors (added Telecom: 26 companies, Education: 31 companies). Same pattern: SEC EDGAR submissions API, USASpending POST search, Senate LDA paginated GET with
page_size=25. Sequential execution only, SQLite locks are still unforgiving.Two uvicorn processes, a scheduler, a Twitter bot cron, nginx, certbot. Still $3.99/month.
Comparison
Same as before. The new engineering is the dual-service architecture and the cross-database evidence source plugin pattern.
Source: https://github.com/Obelus-Labs-LLC/WeThePeople
Second service: https://github.com/Obelus-Labs-LLC/Veritas
lewd_peaches@reddit
For anyone working with larger datasets or computationally intensive tasks, I've found significant speedups by offloading parts of my Python code to GPUs. Not just for ML, but also for things like complex simulations.
I've primarily used PyTorch and CuPy. CuPy is a drop-in replacement for NumPy in many cases, and the performance gains can be substantial. For example, a recent Monte Carlo simulation I was running went from taking 3 hours on my CPU to about 20 minutes on a single RTX 3090. The code change was minimal.
I've also experimented with distributed GPU processing using OpenClaw. I used it to fine-tune a smaller LLM on a dataset that was too large to fit on a single GPU. Setting up the distributed environment took some time initially, but then I was able to run a fine-tuning job across 4 GPUs, finishing in around 6 hours. The cost for the compute was around $25, which was much cheaper than renting a large instance from AWS or GCP. Worth looking into if you're hitting memory limits or need to accelerate your workloads.
kesor@reddit
tmux-player-ctl.py - a controller for MPRIS media players (spotifyd, mpv, mpd, vlc, chrome, ...)
Built tmux-player-ctl.py, a single-file, pure-Python TUI that pops up inside tmux and gives you full keyboard control over any MPRIS media player (spotifyd, mpv, mpd, VLC, Chrome, Firefox, etc.) using
playerctl.When starting to write it I considered various options like bash, rust, go, etc... but Python was the most suitable for what this needed to do and where it needed to go (most Linux distros have python already).
What worked well on from the Python side:
subprocessmodule — both synchronous calls and asynchronous background processes (I run a metadata follower subprocess that pushes real-time updates without blocking the TUI).playerctlcommands.It’s intentionally tiny and fast: launches in a compact tmux popup (
-w72 -h12), shows live track info + progress bar + color-coded volume, supports seek, shuffle, loop modes, and Tab to switch between running players.Typical one-liner:
GitHub: https://github.com/kesor/tmux-player-ctl
I’d especially love feedback from people who regularly wrangle
subprocess, build CLI/TUI tools, or obsess over testing: any patterns I missed, better ways to handle long-running playerctl followers, or testing gotchas you’ve run into? Especially if you have tips on how to deal with ambiguous-width emoji symbols that have different widths in different fonts.TraditionalPen2468@reddit
**agent-aegis** — runtime security library for AI agents
`pip install agent-aegis`
**What My Project Does**
Adds runtime guardrails to AI agent frameworks with two lines of code:
```python
import aegis
aegis.auto_instrument()
```
This auto-patches LangChain, CrewAI, OpenAI Agents SDK, Anthropic, and 6 more frameworks at runtime — every LLM call and tool invocation passes through prompt injection detection (85+ patterns), PII masking (12 categories), and a full audit trail. All deterministic, no LLM calls, sub-millisecond overhead.
Also includes a YAML policy engine for action-level allow/deny rules, 7 approval handlers (Slack, Discord, Telegram, email, webhook), and Merkle tree audit with cryptographic inclusion proofs.
1 dependency (PyYAML). Based on 24 academic papers.
**Target Audience**
Python developers building AI agents who need security controls without adding cloud services or heavy infrastructure. Production use — 5,200+ tests, mypy --strict, sub-millisecond policy evaluation.
**Comparison**
- vs NeMo Guardrails: NeMo uses Colang DSL and focuses on conversational flow control. Aegis focuses on action-level governance (allow/deny tool calls with parameter validation) and works across agent frameworks, not just chat.
- vs Microsoft Agent Governance Toolkit: AGT is an enterprise platform (7 packages, 5 language SDKs, execution sandboxing). Aegis is the lightweight alternative — single pip install, built-in guardrails, zero external service dependencies. Think SQLite vs PostgreSQL.
- vs Guardrails AI: Guardrails AI validates LLM *outputs* (structured data). Aegis validates agent *actions* (tool calls, API requests) and adds policy-based governance.
GitHub: https://github.com/Acacian/aegis | Playground: https://acacian.github.io/aegis/playground/
DifficultDifficulty@reddit
A Python SDK/CLI to make Ray clusters self-serve for Python devs.
What My Project Does
krayne(link) is a Python library and CLI that wraps the KubeRay operator for creating and managing Ray clusters on Kubernetes. Instead of hand-writing KubeRay YAML manifests, you import Python functions (create_cluster(),scale_cluster(),list_clusters(), etc.) or use thekrayne/ikrayne(interactive TUI) CLI to spin up and manage clusters with sensible defaults.The idea is that if you're already writing Ray workflows in Python, training jobs, serve deployments, distributed preprocessing, the cluster management layer should live in the same language. The SDK is the source of truth, the CLI is a thin Typer wrapper on top of it. Operations are stateless functions that return frozen dataclasses, configuration goes through Pydantic models with YAML override support when you need finer control.
GitHub: https://github.com/roulbac/krayne
Target Audience
ML engineers and researchers who write Ray workflows on Kubernetes. The kind of person who knows what
ray.init()does but doesn't want to become a KubeRay manifest expert just to get their cluster running. Also useful for platform teams who want a programmable layer on top of KubeRay that their users can call from Python. It's early (v0.1.0) and opinionated, a composable starting point, not a production-hardened product.Comparison
An alternative I'm familiar with is using
kubectl applywith raw KubeRay manifests, or the KubeRay Python client directly. The main difference is thatkrayneis designed around progressive disclosure:krayne create my-cluster --gpus-per-worker 1 --workers 2is a complete command.It's not that working with KubeRay directly can't do what
kraynedoes, it absolutely can. But when you primarily write Ray code and just need a cluster up with the right resources, context-switching into YAML manifests andkubectlis friction you don't need. A typed Python API that validates your input before it hits the cluster and lives right next to your actual Ray code, that's ultimately why I built it.BenjyWiener@reddit
ZyncIO: dual sync/async interfaces with minimal duplication
PyPI GitHub
I recently released ZyncIO, a production-ready, typing-first, lightweight library that lets you write code for both sync and async Python, without code duplication.
Unlike other libraries, ZyncIO doesn’t use threads, nor does it rely on AST transformations. Instead, it’s built on the concept of “sync coroutines.” Basically, if a an `async def` function never actually yields to the event loop, it can be executed without one.
ZyncIO allows libraries that perform IO but don’t themselves need concurrency (such as REST clients) to be provide both a “normal” (sync) API as well as an async API.
You can read the README and source code for more details, but here’s an example:
``` class BaseClient(zyncio.ZyncBase): def init(self, sock: socket.socket) -> None: self.sock: socket.socket = sock
IMPORTANT: The mixins need to come before the base class!
class SyncClient(zyncio.SyncMixin, BaseClient): pass
class AsyncClient(zyncio.AsyncMixin, BaseClient): def init(self, sock: socket.socket) -> None: super().init(sock) self.sock.setblocking(False)
sync_client = SyncClient(sock) sync_client.do_handshake() # Magically sync! print('Status:', sync_client.status) # Sync property
async def use_async_client(): async_client = AsyncClient(sock) await async_client.do_handshake() # Magically async! print('Status:', await sync_client.status()) # Async func
asyncio.run(use_async_client()) ```
Typing
ZyncIO is fully typed, and designed around Python’s type system to provide seamless IDE/code completion support.
Production-ready
ZyncIO is not a toy project, it’s already used in production.
If you want to see an example of ZyncIO in action, take a look at https://github.com/doronz88/rpc-project, which just gained async support via ZyncIO.
macjaf@reddit
tokencap - a Python library for token budget enforcement across AI agents.
The problem: provider spending caps are account-level and reactive. They tell you what happened after the fact. tokencap enforces limits in your code, before the call goes out.
Two ways to use it:
Direct SDK:
client = tokencap.wrap(anthropic.Anthropic(), limit=50_000)Any agent framework (LangChain, CrewAI, AutoGen, LlamaIndex):
tokencap.patch(limit=50_000)Four actions at configurable thresholds: WARN, DEGRADE (transparent model swap to a cheaper model), BLOCK, and WEBHOOK. SQLite out of the box, Redis for multi-agent setups. Tracks tokens not dollars — token counts come directly from the provider response and never drift with pricing changes.
pip install tokencaphttps://github.com/pykul/tokencap
cwt114@reddit
7 months ago, I shared NeoSQLite v1.0.0 here. It was a simple idea: Give SQLite a PyMongo API so we can have the NoSQL experience in Python without the "NoSQL Server" overhead.
The feedback was amazing (and admittedly, a bit brutal). You guys rightly pointed out the flaws and edge cases. So, I went back to the lab. 374 commits later, it's no longer just a "wrapper" falling back to Python loops—it's a full-blown database engine.
What My Project Does
NeoSQLite gives you the complete NoSQL/MongoDB experience in Python without the infrastructure overhead. It turns a standard SQLite database into a MongoDB-compatible engine.
For Python apps, it's a completely serverless, in-process library. But for this release, I also built the "Magic trick": NX-27017, an optional (and permanently experimental) tiny daemon that speaks the actual MongoDB wire protocol. You can point any existing project, GUI tool like MongoDB Compass, or non-Python app at a single SQLite file with zero code changes.
Target Audience
This is meant for production use in specific contexts: desktop apps, CLI tools, local development environments, IoT devices, and small-to-medium backend services.
If you are building a massive, horizontally scaled enterprise cluster, use a real server. But if you want a drop-in PyMongo replacement that lives in a single file, this is for you.
I know replacing your database engine sounds terrifying, so to sleep at night, I've built a testing suite of 2,600+ unit tests and an automated "compatibility lab". It runs 377 different complex scenarios against both NeoSQLite and a real MongoDB instance to assert the results are strictly identical. We are sitting at 100% API parity for all comparable features.
Real-World Usage
It's actually being used out in the wild now! For example, Andy Felong recently wrote a full blog post about using NeoSQLite for his astronomy projects across a Raspberry Pi Zero, a headless Ubuntu server, and a Mac:
Comparison
Making it "Production Fast"
In the early days, complex queries were slow because I was evaluating them in Python. I've spent the last few months pushing that logic down into raw SQL:
Try it:
pip install neosqliteGitHub: https://github.com/cwt/neosqlite
I'd love to hear your thoughts. Roast me again, or tell me what feature is keeping you tied to a "real" database server for local dev!
The Boring Stats for those interested: 374 commits since v1.0.0, 460 files changed (+105k lines), 30+ releases.
ResponsibleBuilder67@reddit
SDX (Stable Diffusion Extreme) – A Clean, Modern Diffusion Transformer Framework Built with the Latest Techniques and Targeted Fixes
Key Features:
The framework is extensively documented, including detailed explanations of the generation pipeline and the Holy Grail sampler.
GitHub: https://github.com/Llunarstack/sdx
If you're working with Diffusion Transformers or exploring Flow Matching and advanced sampling techniques, SDX provides a solid, forward-looking foundation.
Note: The model hasn't been trained; therefore, there are no results to verify its superiority over current models.
AssociateEmotional11@reddit
Project Name: PyNeat (Upcoming v2.0)
What it does: An AST-based auto-fixer specifically designed to clean up the exact "AI slop" mentioned in this thread's description.
Standard formatters like Black or Ruff are great for styling, but they don't fix bad structural logic. PyNeat uses Instagram's
LibCSTto safely rewrite the AST while preserving 100% of your original comments and whitespace.Currently building v2.0 which targets AI-generated artifacts:
print()statements, JS artifacts likeconsole.log, and useless AI boilerplate comments (# Generated by AI, empty# TODO:).if(arrow anti-patterns) into guard clauses and removes LLM tautologies (e.g., convertingif var == True:->if var:).except: passorprint(e)with saferaise NotImplementedErrorstubs.Status: Just passed massive integration stress-tests against the Anthropic SDK and Pydantic core without breaking the AST. Currently finalizing batch processing (
pyproject.tomlsupport) before the official release.Question for the thread: What is the most annoying "AI coding habit/artifact" you constantly find yourself fixing manually? I'd love to add a rule for it before launching!
nicholashairs@reddit
Big-Rent1128@reddit
RPGNLP, a Python package that tokenizes raw user input for RPG games
Background:
I began working on this package earlier this year when I was making a text-based RPG game. I realized that tokenizing and extracting relevant information from raw text input was more of an undertaking than I thought. So I built an NLP engine on top of NLTK and spaCy to give developers a way to turn raw text into actionable tokens.
What the Project Does:
The engine will take text like "attack the goblin with the hammer" and output a dictionary with values like action: attack, subject: goblin, instrument: hammer. Or "go south east" will output action: travel, direction: south east.
The verbs that the user types is converted into a canon action to make it easier for a game engine to use the data. For instance, if the user types "go south" or "head south," they both tokenize as a "travel" action.
Comparison and Target Audience:
Unlike other NLP packages, this one is specifically designed for RPG games. Hopefully game developers can find this useful so they do not have to develop this sort of engine on their own.
ZyF69@reddit
I've released a new version of Makrell,
v0.10.0. Makrell was originally for the Python platform only, but has expanded into a family of programming languages and tools for metaprogramming, code generation, and language-oriented programming on multiple platforms. I still consider it alpha, so expect errors and missing bits and pieces, but there's a lot of ground covered now. This release includes:The stuff is at https://makrell.dev . For an in-depth introduction, go straight to the article at https://makrell.dev/odds-and-ends/makrell-design-article.html
An AI usage declaration:
Done by me: All language design, MakrellPy, the MakrellPy bits in VS Code extension and the MakrellPy LSP, sample code, basic documentation.
Done by coding agents: Porting to Makrell# and MakrellTS, the MRDT format implementations, the VS Code extension bits for those tracks, the LSP work for those tracks, a lot of documentation, MakrellTS playground, a lot of testing and refinements, packaging. (It was awesome, by the way.)
The coding agent story is a bit special to me. Earlier this year I had to retire after 30 years as a software developer. Due to Parkinson's disease I suffer from fatigue and fine motor control issues that make it hard to do a lot of coding, or regular work at all. Luckily, my congnitive abilities are still good, though. This ironically coincided with the rise of AI coding assistants, which means I can still produce a lot of code while concentrating on design and high-level directions. The Makrell project had been dormant for two years, but now I was suddenly able to make a lot of progress again by using coding agents to do the actual coding work under my direction. I think it's great. I can concentrate on the interesting bits and not spend my limited energy on the more mechanical coding work. Which really isn't that interesting, I should say.
Now the question is if anyone is going to use or care about this. Probably not. And I believe the future of coding is agents compiling directly from specs to machine code and other low level targets, and that few will care about our beatiful programming languages. Maybe I'll just submit this somewhere as a piece of conceptual art.
Below is a blurb meant for language design people.
About Makrell
Makrell is a structural language family built around a shared core called MBF: a bracket-and-operator-based format meant to support code, data, markup, and embedded DSLs without treating them as completely separate worlds. The project currently includes three host-language tracks, MakrellPy, MakrellTS, and Makrell#, plus related formats: MRON for structured data, MRML for markup, and MRTD for typed tabular data.
What may be most interesting to PL people is that Makrell is not being treated as “one syntax, one implementation”. The same family ideas are being pushed through Python, TypeScript/browser, and .NET/CLR hosts, with a specs-first approach and explicit parity work between the tracks. The aim is not to force every host into identical behaviour everywhere, but to separate what belongs to the shared family core from what should remain host-shaped.
The language side has real macro and compile-time machinery rather than just surface syntax sugar. Makrell supports quoting/unquoting, structural rewrites,
meta, and small embedded sublanguages. One of the nicer recurring examples is a shared macro showcase where the same family-level ideas are expressed across the implementations: pipeline reshaping, postfix-to-AST rewriting, and a Lisp-like nested notation living inside Makrell. That general “languages inside languages” direction is a big part of the project’s identity.The formats are not side projects bolted on afterwards. MRON, MRML, and MRTD are meant to demonstrate that the same structural basis can also support data and document-like representations. So Makrell is partly a programming-language project, partly a language-workbench experiment, and partly an attempt to make code, markup, and structured data feel more closely related than they usually do.
v0.10.0is the first release where the whole thing feels like a coherent public system rather than a pile of experiments. The packages are published, the .NET CLI ships as a real tool, the TypeScript track has a standalone browser playground, the VS Code extension covers the three language tracks plus the family formats, and the docs/release story are much more consolidated. The editor path is especially important now: run/check workflows and diagnostics exist across MakrellPy, MakrellTS, Makrell#, MRON, MRML, and MRTD, with a longer-term plan to converge tooling further around a TypeScript-based family language-server direction.If you are interested in macro systems, multi-host language design, little languages, structural notations, or the boundary between programming language and data/markup language design, that is the niche Makrell is trying to explore. It is not “a better Python” or “a replacement for TypeScript”; it is much more a family-oriented design project that happens to have serious implementations in those ecosystems.
The practical entry points now are:
makrell.devfor the overall language-family/docs storyvscode-makrellfor the current editor workflowThe repo still contains a lot of active design work, but
v0.10.0is meant to be the point where the project becomes legible as a real language-family effort instead of only an internal exploration.sharik709@reddit
What My Project Does
lucid-pipelinelets you chain data through a series of processing steps in a clean, readable way. Instead of deeply nested function calls or clunky for loops, you describe the steps declaratively.Instead of:
You write:
It supports plain functions, class-based pipes with a
next_pipecallback (so you can run logic before AND after downstream steps — like middleware), conditional steps with.when()/.unless(), side effects with.tap(), error handling with.on_failure(), and a full async version.Class-based pipes can also short-circuit by not calling
next_pipe:Zero dependencies. \~14KB. Python 3.10+.
Target Audience
This is meant for production use. Anyone doing multi-step data processing — ETL scripts, request/response middleware, form validation chains, content transformation pipelines. Anywhere you find yourself writing sequential transformations with conditional logic and error handling mixed in.
I've been using it at work for cleaning up request validation and ETL jobs. It's on PyPI as v0.1.0 and fully tested (sync and async).
Comparison
lucid-pipelineis simpler — it's about clean sequential composition, not parallelism.|operator for functional-style chaining. Clean for simple transforms, but doesn't support middleware-style wrapping (before/after logic), conditional steps, or error handling built in.lucid-pipelineis for in-process, in-memory step chaining — no infrastructure, no DAGs, no config files. Think "inside a single function" not "across a cluster."Links:
pip install lucid-pipelineHappy to hear feedback on the API design. This is v0.1.0 and the first package in a small ecosystem I'm building around bringing more convention to Python projects.
Candid_Complaint_925@reddit
BillingWatch — Self-Hosted Stripe Billing Anomaly Detector
Built this because Baremetrics/ProfitWell felt overkill for solo devs who just want to know when something's wrong with their Stripe payments.
FastAPI app that processes Stripe webhooks in real-time and flags anomalies — unexpected refunds, payment failure spikes, revenue drops. Dashboard shows per-tenant billing health. No cloud required, you own your data.
Quick start:
One-click deploy configs included for Railway/Render/Fly. MIT licensed.
Repo: https://github.com/rmbell09-lang/BillingWatch
Ok_Leading4235@reddit
aiofastnet - optimized (up to x2.2 faster) drop-in replacements for asyncio networking APIs
As a part of algorithmic trading project I had to look into the actual performance of uvloop and asyncio network API. Turned out it wasn't so great, TLS part is especially bad, also in uvloop. Lot's of plumbing code and memory copying. I tried to push PRs to uvloop but the project is almost unmaintained these days. Took more than 1 year to get some of the PRs reviewed and merged.
Eventually I came up with a much cleaner and loop agnostic way to improve networking API performance.
https://github.com/tarasko/aiofastnet
What My Project Does
Provides optimized versions of asyncio networking APIs:
Target Audience
This project is mainly for developers who already use asyncio transports/protocols and want better performance without redesigning their code.
It is probably most relevant for people building:
Comparison
Compared to uvloop/winloop, aiofastnet is not a separate event loop. It focuses specifically on the transport/TLS layer and works with the loop you already use.
Feedback is very welcome!
Due_Anything4678@reddit
ghostdep - finds phantom and unused deps in your Python project
What My Project Does
Scans your Python project and tells you what you import but didn't add to your manifest, and what you declared but never use.
$ ghostdep -p my-project[phantom] pandas atapp.py:7[unused] numpy at requirements.txtHandles requirements.txt, pyproject.toml (PEP 621, Poetry, uv/PEP 735). Knows about aliases like PIL→Pillow, cv2→opencv-python, sklearn→scikit-learn. Uses tree-sitter for AST parsing, not regex.
Single binary, no Python runtime needed. Also supports Go, JS/TS, Rust, Java if you work across languages.
cargo install ghostdephttps://github.com/ojuschugh1/ghostdep
Target Audience
Anyone maintaining Python projects who wants cleaner dependency manifests. Works in CI too - has JSON and SARIF output, exit code 1 when findings exist. v0.1.0, looking for feedback.
Comparison
Most Python dep checkers (pip-check, pip-audit, safety) focus on vulnerabilities or version conflicts. ghostdep focuses on a different problem: deps that are imported but not declared (phantom) or declared but never imported (unused). Closest tool is probably deptry - ghostdep differs by being cross-language (5 languages in one binary) and using AST parsing with confidence scoring for dynamic/conditional imports.
Powerful_Lock6120@reddit
mpv-trackeris a Textual TUI for tracking local anime / series watched inmpv, with MyAnimeList integration built in.Features:
Install:
uvx mpv-trackerpipx install mpv-trackerpip install mpv-trackerLinks:
Demo: https://github.com/GenessyX/mpv-tracker?tab=readme-ov-file#showcase
Built this for a local anime +
mpvworkflow where I wanted something lighter than a full media manager, but still with MAL sync and a usable TUI.Candid_Complaint_925@reddit
TradeSight — Self-Hosted Strategy Backtesting + Paper Trading Lab
Self-hosted Python app that runs strategy backtests, paper trades via Alpaca, and lets you run overnight strategy tournaments (9 strategies competing, best allocation wins). All local, no cloud subscription.
Quick start:
Runs in demo mode without any API keys. Strategies: MACD, RSI, Bollinger, SMA/EMA crossovers, ATR-based. 169/169 tests passing.
Wrote a comparison of Python backtesting options if that's useful: https://gist.github.com/rmbell09-lang/129d53d4c471eafa4f71d19a4c168a72
Repo: https://github.com/rmbell09-lang/tradesight — MIT licensed, 33 stars.
Chunky_cold_mandala@reddit
GitGalaxy- A hyper-scale static analyzer & threat-hunting engine built on DNA sequencing principles
What my project does -
GitGalaxy is a two-part ecosystem. It is designed to extract the structural DNA of massive software repositories and render their non-visual architecture into measurable, explorable 3D galaxies.
1. The blAST Engine - The galaxyscope (Backend): A hyper-scale, language-agnostic static analysis CLI. Based on 50 years of bioinformatics and genetic sequencing algorithms, it parses code at \~100,000 LOC/second. It outputs rich JSON telemetry, SQLite databases, and low-token Markdown briefs optimized for AI-agent workflows.
2. The Observatory (Frontend): Drop your galaxy.json into the free viewer at GitGalaxy.io or use the repo's airgap_observatory, a standalone, zero-telemetry WebGPU visualizer. Both visualizers read the JSON contract and renders the entire code base as a procedural 3D galaxy where files are stars, allowing humans to visually map scale and risk exposure instantly.
Live Demo: View 3D galaxy examples of Apollo-11, Linux, Tensorflow and more at GitGalaxy.io - - github - https://github.com/squid-protocol/gitgalaxy
The blAST Paradigm: Sequencing the DNA of Software
Traditional computer science treats software like a rigid blueprint, using slow, language-specific Abstract Syntax Trees (ASTs) to analyze code. GitGalaxy treats code as a sequence to be scanned and then analyzed for patterns and occurrences using the blAST (Broad Lexical Abstract Syntax Tracker) engine.
By applying the principles of biological sequence alignment to software, blAST hunts for the universal structural markers of logic across \~40 languages and \~250 file extensions. We translate this genetic code into "phenotypes"—measurable risk exposures.
Sequencing at Hyper-Scale
By abandoning the compiler bottleneck, blAST achieves processing velocities that traditional ASTs simply cannot comprehend. In live telemetry tracking across the largest open-source ecosystems, blAST demonstrated its absolute scale:
Zero-Trust Architecture
Your code never leaves your machine. GitGalaxy performs 100% of its scanning and vectorization locally.
The Viral Security Lens: Behavioral Threat Hunting
Traditional security scanners rely on rigid, outdated virus signatures. blAST acts like an immune system, hunting for the behavioral genetic markers of a threat. By analyzing the structural density of I/O hits, execution triggers, and security bypasses, blAST is perfectly engineered to stop modern attack vectors:
eval/exec)..pem,.pfx,.jksfiles) buried deep within massive repositories.