# Appendix B — Open-Source Libraries and Engineering Tools
This appendix lists curated open-source tools for neuro-symbolic algorithms, knowledge-graph construction, and cloud–edge deployment. For Appendix C labs or a low-altitude governance prototype, you can compose a toolchain from the sections below.

### B.1 PyKEEN
**PyKEEN (Python KnowledgE EmbeddiNgs)** is a leading Python library for training and evaluating knowledge-graph embedding (KGE) models.
* **Core capabilities:** Dozens of classic KGE models (TransE, RotatE, ComplEx, …) with standard training pipelines and metrics (MRR, Hits@K).
* **Use in this book:** Chapters 8–9 embed static SkyKG topology (e.g., airspace adjacency, route connectivity) into dense vectors as priors for downstream deep models.

### B.2 DeepProbLog
**DeepProbLog** is a prominent probabilistic logic programming framework for neuro-symbolic AI, extending ProbLog with neural components.
* **Core capabilities:** Neural networks appear as **neural predicates** in first-order rules; the engine performs exact logical inference and backpropagation over hybrid programs.
* **Use in this book:** Fits Layer 1/2 hybrid systems (Chapter 7). Vision/radar classifiers can be neural predicates coupled with aviation regulations as symbolic predicates for joint training.

### B.3 Logic Tensor Networks (LTN)
**Logic Tensor Networks (LTN)** fuse real-valued (fuzzy) logic with deep learning computation graphs.
* **Core capabilities:** T-norms turn Boolean formulas into differentiable losses; TensorFlow or PyTorch backends supported.
* **Use in this book:** Primary practice vehicle for “logic as loss” (Chapter 8). Encode CCAR-style separation constraints as formulas penalizing trajectory predictors during training.

### B.4 PyTorch Geometric (PyG)
**PyTorch Geometric** is the dominant PyTorch extension for graph neural networks.
* **Core capabilities:** Sparse/dynamic graph kernels, GCN/GAT/GraphSAGE-style message passing, temporal and heterogeneous graph utilities.
* **Use in this book:** Part IV (dynamic coordination)—foundation for Chapter 12’s graph-driven multi-agent conflict models; temporal attention on high-rate TKG snapshots for millisecond-scale sensing.

### B.5 RDFLib / OWLReady2 / Neo4j / GraphDB
These tools cover the KG lifecycle from ontology modeling to graph storage.
* **RDFLib & OWLReady2:** Native Python semantic-web stacks. RDFLib reads/writes/queries RDF; OWLReady2 loads and reasons over OWL ontologies in Python—useful for SkyKG concept and rule layers (Chapter 6).
* **Neo4j:** Industrial property-graph database with Cypher—fast traversals for dynamic traffic workloads.
* **GraphDB:** Enterprise RDF/SPARQL store with forward-chaining inference (OWL 2 RL/QL fragments)—a natural backend for deterministic static compliance checks (Chapter 10, rule-query channel).

### B.6 LLM + KG Integration Stacks
Middleware connecting large models to structured knowledge is central to GraphRAG (Chapters 10 and 22).
* **LangChain & LlamaIndex:** Mainstream LLM orchestration frameworks with loaders for graph stores, SPARQL/Cypher generation, and multi-hop subgraph retrieval.
* **Microsoft GraphRAG:** Microsoft’s open KG-RAG stack emphasizing hierarchical entities and global summaries—helpful for fleet-level situational questions beyond local chunk retrieval.

### B.7 Streaming Inference and Edge Deployment
To bridge the compute–latency gap (Chapter 18) and deploy in the field, use stream processing and edge runtimes.
* **Apache Flink / Kafka:** High-throughput streaming and messaging—ingest ~10 Hz ADS-B feeds and drive event-driven TKG updates.
* **TensorRT & ONNX Runtime:** Quantization and optimized inference—deploy trained GNNs and light perception nets to 5G MEC or onboard Jetson-class devices for low-latency interventions.
