A Framework for Building Context Graphs and Decision Intelligence Layers for AI
Transform Chaos into Intelligence. Build AI systems with context graphs, decision tracking, and advanced knowledge engineering that are explainable, traceable, and trustworthy β not black boxes.
AI agents today are capable but not trustworthy:
- No memory structure β agents store embeddings, not meaning. Retrieval is fuzzy; there's no way to ask why something was recalled.
- No decision trail β agents make decisions continuously but record nothing. When something goes wrong, there's no history to debug or audit.
- No provenance β outputs cannot be traced back to source facts. In regulated industries, this is a compliance blocker.
- No reasoning transparency β black-box answers with no explanation of how a conclusion was reached.
- No conflict detection β contradictory facts silently coexist in vector stores, producing unpredictable answers.
These aren't edge cases. They are the reason AI cannot be deployed in healthcare, finance, legal, and government without custom guardrails built from scratch.
Semantica is the context and intelligence layer you add to your AI stack:
- Context Graphs β structured graph of entities, relationships, and decisions your agent builds as it works. Queryable, traceable, persistent.
- Decision Intelligence β every decision is a first-class object: recorded, linked causally, searchable by precedent, and analyzable for downstream impact.
- Provenance β every fact links to its source. W3C PROV-O compliant. Full lineage from ingestion to inference.
- Reasoning engines β forward chaining, Rete networks, deductive, abductive, and SPARQL reasoning. Explainable inference paths, not black-box answers.
- Deduplication & QA β conflict detection, entity resolution, and validation built into the pipeline.
Works alongside LangChain, LlamaIndex, AutoGen, CrewAI, and any LLM provider β Semantica is not a replacement, it's the accountability layer on top.
pip install semanticaFirst stable release β
Production/Stableon PyPI. Ships across three stages: 0.3.0-alpha, 0.3.0-beta, and 0.3.0 stable.
| Area | Highlights |
|---|---|
| Context Graphs | Temporal validity windows (valid_from/valid_until), weighted BFS (min_weight), cross-graph navigation (link_graph, navigate_to, resolve_links) with full save/load persistence |
| Decision Intelligence | Complete lifecycle: record_decision β trace_decision_chain β analyze_decision_impact β find_similar_decisions; hybrid precedent search; PolicyEngine with versioned rules |
| KG Algorithms | PageRank, betweenness, community detection (Louvain), Node2Vec embeddings, link prediction, path finding β all returning structured dicts |
| Semantic Extraction | LLM relation extraction fixed (no silent drops); _match_pattern rewritten; duplicate relation bug removed; "llm_typed" metadata corrected |
| Deduplication v2 | blocking_v2/hybrid_v2 candidate generation (63.6% faster); two-stage prefilter (18β25% faster); semantic dedup v2 (6.98x faster) |
| Delta Processing | SPARQL-based incremental diff; delta_mode pipelines; snapshot versioning with prune_versions() |
| Export | RDF format aliases ("ttl", "json-ld", etc.); ArangoDB AQL export; Apache Parquet export (Spark/BigQuery/Databricks ready) |
| Pipeline | FailureHandler with LINEAR/EXPONENTIAL/FIXED backoff; PipelineValidator returning ValidationResult; retry loop fixed |
| Graph Backends | Apache AGE (SQL injection fixed), AWS Neptune, FalkorDB, PgVector (HNSW/IVFFlat indexing) |
| Tests | 886+ passing, 0 failures β 335 context, ~430 KG, 70 semantic extraction, 85 real-world E2E |
See RELEASE_NOTES.md for the full per-contributor breakdown and CHANGELOG for the complete diff.
- Context Graphs β structured graph of entities, relationships, and decisions; queryable, causal, persistent
- Decision tracking β record, link, and analyze every agent decision with
add_decision(),record_decision() - Causal chains β link decisions with
add_causal_relationship(), trace lineage withtrace_decision_chain() - Precedent search β hybrid similarity search over past decisions with
find_similar_decisions() - Influence analysis β
analyze_decision_impact(),analyze_decision_influence()β understand downstream effects - Policy engine β enforce business rules with
check_decision_rules(); automated compliance validation - Agent memory β
AgentMemorywith short/long-term storage, conversation history, and statistics - Cross-system context capture β
capture_cross_system_inputs()for multi-agent pipelines
- Knowledge graph construction β entities, relationships, properties, typed edges
- Graph algorithms β PageRank, betweenness centrality, clustering coefficient, community detection
- Node embeddings β Node2Vec embeddings via
NodeEmbedder - Similarity β cosine similarity via
SimilarityCalculator - Link prediction β score potential new edges via
LinkPredictor - Temporal graphs β time-aware nodes and edges
- Incremental / delta processing β update graphs without full recompute
- Entity extraction β named entity recognition, normalization, classification
- Relation extraction β triplet generation from raw text using LLMs or rule-based methods
- LLM-typed extraction β extraction with typed relation metadata
- Deduplication v1 β Jaro-Winkler similarity, basic blocking
- Deduplication v2 β
blocking_v2,hybrid_v2,semantic_v2strategies withmax_candidates_per_entity - Triplet deduplication β
dedup_triplets()for removing duplicate (subject, predicate, object) triples
- Forward chaining β
Reasonerwith IF/THEN string rules and dict facts - Rete network β
ReteEnginefor high-throughput production rule matching - Deductive reasoning β
DeductiveReasonerfor classical inference - Abductive reasoning β
AbductiveReasonerfor hypothesis generation from observations - SPARQL reasoning β
SPARQLReasonerfor query-based inference over RDF graphs
- Entity provenance β
ProvenanceTracker.track_entity(id, source_url, metadata) - Algorithm provenance β
AlgorithmTrackerWithProvenancetracks computation lineage - Graph builder provenance β
GraphBuilderWithProvenancerecords entity source lineage from URLs - W3C PROV-O compliant β lineage tracking across all modules
- Change management β version control with checksums, audit trails, compliance support
- Backends β FAISS, Pinecone, Weaviate, Qdrant, Milvus, PgVector, in-memory
- Semantic search β top-k retrieval by embedding similarity
- Hybrid search β vector + keyword with configurable weights
- Filtered search β metadata-based filtering on any field
- Custom similarity weights β tune retrieval per use case
- AWS Neptune β Amazon Neptune graph database with IAM authentication
- Apache AGE β PostgreSQL graph extension with openCypher via SQL
- FalkorDB β native support;
DecisionQueryandCausalChainAnalyzerwork directly with FalkorDB row/header shapes
- File formats β PDF, DOCX, HTML, JSON, CSV, Excel, PPTX, archives
- Web crawl β
WebIngestorwith configurable depth - Databases β
DBIngestorwith SQL query support - Snowflake β
SnowflakeIngestorwith table/query ingestion, pagination, and key-pair/OAuth auth - Docling β advanced document parsing with table and layout extraction (PDF, DOCX, PPTX, XLSX)
- Media β image OCR, audio/video metadata extraction
- RDF β Turtle (
.ttl), JSON-LD, N-Triples (.nt), XML viaRDFExporter - Parquet β
ParquetExporterfor entities, relationships, and full KG export - ArangoDB AQL β ready-to-run INSERT statements via
ArangoAQLExporter - OWL ontologies β export generated ontologies in Turtle or RDF/XML
- Pipeline builder β
PipelineBuilderwith stage chaining and parallel workers - Validation β
PipelineValidatorreturnsValidationResult(valid, errors, warnings)before execution - Failure handling β
FailureHandlerwithRetryPolicyandRetryStrategy(exponential backoff, fixed, etc.) - Parallel processing β configurable worker count per pipeline stage
- LLM providers β 100+ models via LiteLLM (OpenAI, Anthropic, Cohere, Mistral, Ollama, and more)
- Auto-generation β derive OWL ontologies from knowledge graphs via
OntologyGenerator - Import β load existing OWL, RDF, Turtle, JSON-LD ontologies via
OntologyImporter - Validation β HermiT/Pellet compatible consistency checking
| Module | What it provides |
|---|---|
semantica.context |
Context graphs, agent memory, decision tracking, causal analysis, precedent search, policy engine |
semantica.kg |
Knowledge graph construction, graph algorithms, centrality, community detection, embeddings, link prediction, provenance |
semantica.semantic_extract |
NER, relation extraction, event extraction, coreference, triplet generation, LLM-enhanced extraction |
semantica.reasoning |
Forward chaining, Rete network, deductive, abductive, SPARQL reasoning, explanation generation |
semantica.vector_store |
FAISS, Pinecone, Weaviate, Qdrant, Milvus, PgVector, in-memory; hybrid & filtered search |
semantica.export |
RDF (Turtle/JSON-LD/N-Triples/XML), Parquet, ArangoDB AQL, CSV, YAML, OWL, graph formats |
semantica.ingest |
Files (PDF, DOCX, CSV, HTML), web crawl, feeds, databases, Snowflake, MCP, email, repositories |
semantica.ontology |
Auto-generation (6-stage pipeline), OWL/RDF export, import (OWL/RDF/Turtle/JSON-LD), validation, versioning |
semantica.pipeline |
Pipeline DSL, parallel workers, validation, retry policies, failure handling, resource scheduling |
semantica.graph_store |
Graph database backends β Neo4j, FalkorDB, Apache AGE, Amazon Neptune; Cypher queries |
semantica.embeddings |
Text embedding generation β Sentence-Transformers, FastEmbed, OpenAI, BGE; similarity calculation |
semantica.deduplication |
Entity deduplication, similarity scoring, merging, clustering; blocking and semantic strategies |
semantica.provenance |
W3C PROV-O compliant end-to-end lineage tracking, source attribution, audit trails |
semantica.parse |
Document parsing β PDF, DOCX, PPTX, HTML, code, email, structured data, media with OCR |
semantica.split |
Document chunking β recursive, semantic, entity-aware, relation-aware, graph-based, ontology-aware |
semantica.normalize |
Data normalization for text, entities, dates, numbers, quantities, languages, encodings |
semantica.conflicts |
Multi-source conflict detection (value, type, relationship, temporal, logical) with resolution strategies |
semantica.change_management |
Version storage, change tracking, checksums, audit trails, compliance support for KGs and ontologies |
semantica.triplet_store |
RDF triplet store integration β Blazegraph, Jena, RDF4J; SPARQL queries and bulk loading |
semantica.visualization |
Interactive and static visualization of KGs, ontologies, embeddings, analytics, and temporal graphs |
semantica.seed |
Seed data management for initial KG construction from CSV, JSON, databases, and APIs |
semantica.core |
Framework orchestration, configuration management, knowledge base construction, plugin system |
semantica.llms |
LLM provider integrations β Groq, OpenAI, HuggingFace, LiteLLM |
semantica.utils |
Shared utilities β logging, validation, exception handling, constants, types, progress tracking |
import semantica
from semantica.context import AgentContext, ContextGraph
from semantica.vector_store import VectorStore
# Build an agent with structured context
context = AgentContext(
vector_store=VectorStore(backend="faiss", dimension=768),
knowledge_graph=ContextGraph(advanced_analytics=True),
decision_tracking=True,
kg_algorithms=True,
)
# Store memory
memory_id = context.store(
"GPT-4 outperforms GPT-3.5 on reasoning benchmarks by 40%",
conversation_id="research_session_1",
)
# Record a decision with full context
decision_id = context.record_decision(
category="model_selection",
scenario="Choose LLM for production reasoning pipeline",
reasoning="GPT-4 benchmark advantage justifies 3x cost increase",
outcome="selected_gpt4",
confidence=0.91,
entities=["gpt4", "gpt35", "reasoning_pipeline"],
)
# Find similar decisions from history
precedents = context.find_precedents("model selection reasoning", limit=5)
# Analyze downstream influence of this decision
influence = context.analyze_decision_influence(decision_id)π Full Quick Start β’ π³ Cookbook Examples β’ π¬ Join Discord β’ β Star Us
| Trustworthy | Explainable | Auditable |
|---|---|---|
| Conflict detection & validation | Transparent reasoning paths | Complete provenance tracking |
| Rule-based governance | Entity relationships & ontologies | W3C PROV-O compliant lineage |
| Production-grade QA | Multi-hop graph reasoning | Source tracking & integrity verification |
Semantica complements LangChain, LlamaIndex, AutoGen, CrewAI, Google ADK, Agno, and other frameworks to enhance your agents with:
| Feature | Benefit |
|---|---|
| Context Graphs | Structured knowledge representation with entity relationships and semantic context |
| Decision Tracking | Complete decision lifecycle management with precedent search and causal analysis |
| KG Algorithms | Advanced graph analytics including centrality, community detection, and embeddings |
| Vector Store Integration | Hybrid search with custom similarity weights and advanced filtering |
| Auditable | Complete provenance tracking with W3C PROV-O compliance |
| Explainable | Transparent reasoning paths with entity relationships |
| Provenance-Aware | End-to-end lineage from documents to responses |
| Validated | Built-in conflict detection, deduplication, QA |
| Governed | Rule-based validation and semantic consistency |
| Version Control | Enterprise-grade change management with integrity verification |
| π₯ Healthcare | π° Finance | βοΈ Legal |
|---|---|---|
| Clinical decisions | Fraud detection | Evidence-backed research |
| Drug interactions | Regulatory support | Contract analysis |
| Patient safety | Risk assessment | Case law reasoning |
| π Cybersecurity | ποΈ Government | π Infrastructure | π Autonomous |
|---|---|---|---|
| Threat attribution | Policy decisions | Power grids | Decision logs |
| Incident response | Classified info | Transportation | Safety validation |
- Context Graphs β Structured knowledge representation with entity relationships and semantic context
- Decision Tracking Systems β Complete decision lifecycle management with precedent search and causal analysis
- GraphRAG Systems β Retrieval with graph reasoning and hybrid search using KG algorithms
- AI Agents β Trustworthy, accountable multi-agent systems with semantic memory and decision history
- Reasoning Models β Explainable AI decisions with reasoning paths and influence analysis
- Enterprise AI β Governed, auditable platforms that support compliance and policy enforcement
- Docling Support β Document parsing with table extraction (PDF, DOCX, PPTX, XLSX)
- AWS Neptune β Amazon Neptune graph database support with IAM authentication
- Apache AGE β PostgreSQL graph extension backend (openCypher via SQL)
- Snowflake β Native ingestion with
SnowflakeIngestor; table/query ingestion, pagination, key-pair & OAuth auth - Custom Ontology Import β Import existing ontologies (OWL, RDF, Turtle, JSON-LD)
Built for environments where every answer must be explainable and governed.
Semantica's flagship module. Tracks every decision your agent makes as a structured graph node β with causal links, precedent search, impact analysis, and policy enforcement.
from semantica.context import ContextGraph
graph = ContextGraph(advanced_analytics=True)
# Record a loan approval decision
loan_id = graph.add_decision(
category="loan_approval",
scenario="Mortgage application β 780 credit score, 28% DTI",
reasoning="Strong credit history, stable income for 8 years, low DTI",
outcome="approved",
confidence=0.95,
)
# Record a downstream dependent decision
rate_id = graph.add_decision(
category="interest_rate",
scenario="Set rate for approved mortgage",
reasoning="Prime applicant qualifies for lowest tier rate",
outcome="rate_set_6.2pct",
confidence=0.98,
)
# Link the decisions causally
graph.add_causal_relationship(loan_id, rate_id, relationship_type="enables")
# Find similar past decisions using hybrid similarity
similar = graph.find_similar_decisions("mortgage approval", max_results=5)
chain = graph.trace_decision_chain(loan_id)
impact = graph.analyze_decision_impact(loan_id)
compliance = graph.check_decision_rules({"category": "loan_approval", "confidence": 0.95})
insights = graph.get_decision_insights()from semantica.context import AgentContext, AgentMemory
from semantica.vector_store import VectorStore
context = AgentContext(
vector_store=VectorStore(backend="faiss", dimension=768),
knowledge_graph=ContextGraph(advanced_analytics=True),
decision_tracking=True,
graph_expansion=True,
kg_algorithms=True,
)
context.store("Regulation EU 2024/1689 requires explainability for high-risk AI", conversation_id="compliance_review")
context.store("Our fraud model flags 0.3% of transactions", conversation_id="compliance_review")
results = context.retrieve("AI regulation explainability requirements", limit=3)
history = context.get_conversation_history("compliance_review")
stats = context.get_statistics()from semantica.kg import KnowledgeGraph, Entity, Relationship
from semantica.kg import CentralityAnalyzer, NodeEmbedder, LinkPredictor
kg = KnowledgeGraph()
kg.add_entity(Entity(id="transformer", label="Transformer", type="Architecture",
properties={"year": 2017, "paper": "Attention Is All You Need"}))
kg.add_entity(Entity(id="bert", label="BERT", type="Model",
properties={"year": 2018, "parameters": "340M"}))
kg.add_entity(Entity(id="gpt4", label="GPT-4", type="Model", properties={"year": 2023}))
kg.add_relationship(Relationship(source="bert", target="transformer", type="based_on"))
kg.add_relationship(Relationship(source="gpt4", target="transformer", type="based_on"))
# Graph algorithms
analyzer = CentralityAnalyzer(kg)
centrality = analyzer.compute_pagerank()
betweenness = analyzer.compute_betweenness()
# Node embeddings (Node2Vec)
embedder = NodeEmbedder()
embeddings = embedder.compute_embeddings(kg, node_labels=["Model"], relationship_types=["based_on"])
# Link prediction
predictor = LinkPredictor()
score = predictor.score_link(kg, "gpt4", "bert", method="common_neighbors")
models = kg.find_nodes(type="Model")
descendants = kg.get_neighbors("transformer", direction="incoming")from semantica.semantic_extract import extract_entities, extract_relations, extract_triplets
text = """
OpenAI released GPT-4 in March 2023. Microsoft integrated GPT-4 into Azure OpenAI Service.
Anthropic, founded by former OpenAI researchers, released Claude as a competing model.
"""
entities = extract_entities(text)
# β [Entity(label="OpenAI", type="Organization"), Entity(label="GPT-4", type="Model"), ...]
relations = extract_relations(text)
# β [Relation(source="OpenAI", type="released", target="GPT-4"), ...]
triplets = extract_triplets(text)from semantica.deduplication import DuplicateDetector
entities = [
{"id": "e1", "name": "OpenAI Inc.", "type": "Organization"},
{"id": "e2", "name": "Open AI", "type": "Organization"},
{"id": "e3", "name": "Anthropic", "type": "Organization"},
]
detector = DuplicateDetector()
duplicates = detector.detect_duplicates(entities, threshold=0.85)
# β [("e1", "e2")]
duplicates_v2 = detector.detect_duplicates(entities, threshold=0.85, strategy="semantic_v2")from semantica.reasoning import Reasoner
reasoner = Reasoner()
reasoner.add_rule("IF Person(?x) THEN Mortal(?x)")
reasoner.add_rule("IF Employee(?x) AND WorksAt(?x, ?y) THEN HasEmployer(?x, ?y)")
results = reasoner.infer_facts([
"Person(Socrates)",
"Employee(Alice)",
{"source_name": "Alice", "target_name": "OpenAI", "type": "WorksAt"},
])
# β ["Mortal(Socrates)", "HasEmployer(Alice, OpenAI)"]from semantica.reasoning import ReteEngine
rete = ReteEngine()
rete.add_rule({
"name": "flag_high_risk_transaction",
"conditions": [
{"field": "amount", "operator": ">", "value": 10000},
{"field": "country", "operator": "in", "value": ["IR", "KP", "SY"]},
],
"action": "flag_for_compliance_review",
})
matches = rete.match({"amount": 15000, "country": "IR", "id": "txn_9921"})from semantica.reasoning import DeductiveReasoner, AbductiveReasoner
deductive = DeductiveReasoner()
deductive.add_axiom("All transformers use attention mechanisms")
deductive.add_fact("BERT is a transformer")
conclusion = deductive.reason("Does BERT use attention?")
abductive = AbductiveReasoner()
abductive.add_observation("The model accuracy dropped 12% after deployment")
hypotheses = abductive.generate_hypotheses()
# β ["Distribution shift in production data", "Preprocessing pipeline mismatch", ...]W3C PROV-O compliant lineage tracking. Every fact traces back to its origin.
from semantica.kg import ProvenanceTracker, AlgorithmTrackerWithProvenance
tracker = ProvenanceTracker()
tracker.track_entity("gpt4_benchmark",
source_url="https://openai.com/research/gpt-4",
metadata={"metric": "MMLU", "score": 86.4})
algo_tracker = AlgorithmTrackerWithProvenance(provenance=True)
algo_tracker.track_graph_construction(
algorithm="node2vec",
input_data={"nodes": 1500, "edges": 4200},
parameters={"dimensions": 128, "walk_length": 80},
)
sources = tracker.get_all_sources("gpt4_benchmark")
all_entities = tracker.get_all_entities()from semantica.vector_store import VectorStore
vs = VectorStore(backend="faiss", dimension=768)
vs.store("The Transformer architecture revolutionized NLP",
metadata={"source": "arxiv", "year": 2017}, id="doc_001")
vs.store("BERT introduced bidirectional pre-training for language understanding",
metadata={"source": "arxiv", "year": 2018}, id="doc_002")
results = vs.search("attention mechanisms in language models", top_k=5)
results = vs.hybrid_search(
query="transformer pre-training",
top_k=10,
vector_weight=0.6,
keyword_weight=0.4,
)
results = vs.search("pre-training", top_k=5, filter={"year": 2018})from semantica.ingest import FileIngestor, WebIngestor, DBIngestor
file_ingestor = FileIngestor(recursive=True)
docs = file_ingestor.ingest("./research_papers/")
web_ingestor = WebIngestor(max_depth=2)
web_docs = web_ingestor.ingest("https://arxiv.org/abs/1706.03762")
db_ingestor = DBIngestor(connection_string="postgresql://user:pass@localhost/kg_db")
db_docs = db_ingestor.ingest(query="SELECT title, abstract FROM papers WHERE year >= 2020")
all_sources = docs + web_docs + db_docsfrom semantica.parse import DoclingParser
# Advanced table and layout extraction
docling = DoclingParser()
parsed = docling.parse("financial_report.pdf")from semantica.ingest import SnowflakeIngestor
# Connect to Snowflake and ingest a table
ingestor = SnowflakeIngestor(
account="myorg-myaccount",
user="analyst",
password="...",
warehouse="COMPUTE_WH",
database="ANALYTICS",
schema="PUBLIC",
)
# Ingest a table with optional filtering and pagination
data = ingestor.ingest_table(
table_name="customer_events",
where="event_date >= '2024-01-01'",
limit=10000,
)
# Or run a custom SQL query
data = ingestor.ingest_query(
query="SELECT id, content, tags FROM knowledge_base WHERE active = TRUE",
batch_size=500,
)
# Convert to Semantica documents for downstream pipeline use
docs = ingestor.export_as_documents(data, id_field="id", text_fields=["content"])
# Key-pair and OAuth auth are also supported via env vars:
# SNOWFLAKE_PRIVATE_KEY_PATH, SNOWFLAKE_TOKEN, SNOWFLAKE_AUTHENTICATORfrom semantica.export import RDFExporter, ParquetExporter, ArangoAQLExporter
rdf_exporter = RDFExporter()
turtle = rdf_exporter.export_to_rdf(kg, format="turtle")
jsonld = rdf_exporter.export_to_rdf(kg, format="json-ld")
ntriples = rdf_exporter.export_to_rdf(kg, format="nt")
parquet_exporter = ParquetExporter()
parquet_exporter.export_entities(kg, path="output/entities.parquet")
parquet_exporter.export_relationships(kg, path="output/relationships.parquet")
parquet_exporter.export_knowledge_graph(kg, path="output/")
aql_exporter = ArangoAQLExporter()
aql_exporter.export(kg, path="output/insert.aql")from semantica.pipeline import PipelineBuilder, PipelineValidator, FailureHandler
from semantica.pipeline import RetryPolicy, RetryStrategy
builder = (
PipelineBuilder()
.add_stage("ingest", FileIngestor(recursive=True))
.add_stage("extract", extract_triplets)
.add_stage("deduplicate", DuplicateDetector())
.add_stage("build_kg", KnowledgeGraph())
.add_stage("export", RDFExporter())
.with_parallel_workers(4)
)
validator = PipelineValidator()
result = validator.validate(builder)
if result.valid:
pipeline = builder.build()
pipeline.run(input_path="./documents/")
retry_policy = RetryPolicy(strategy=RetryStrategy.EXPONENTIAL_BACKOFF, max_retries=3)
handler = FailureHandler()
handler.handle_failure(error=last_error, policy=retry_policy, retry_count=1)from semantica.ontology import OntologyGenerator, OntologyImporter
generator = OntologyGenerator()
ontology = generator.generate(kg)
generator.export(ontology, path="domain_ontology.owl", format="turtle")
importer = OntologyImporter()
ontology = importer.load("existing_ontology.owl")
ontology = importer.load("schema.ttl", format="turtle")
ontology = importer.load("context.jsonld")Graph Databases
- AWS Neptune β Amazon Neptune with IAM authentication
- Apache AGE β PostgreSQL + openCypher via SQL
- FalkorDB β native support for decision queries and causal analysis
Vector Databases
- FAISS β high-performance dense vector search
- Pinecone β serverless and pod-based managed vector database (
pip install semantica[vectorstore-pinecone]) - Weaviate β GraphQL-based vector store with rich schema management (
pip install semantica[vectorstore-weaviate]) - Qdrant β collection-based store with payload filtering (
pip install semantica[vectorstore-qdrant]) - Milvus β scalable store with partition support and multiple index types (
pip install semantica[vectorstore-milvus]) - PgVector β PostgreSQL pgvector extension with JSONB metadata (
pip install semantica[vectorstore-pgvector]) - In-memory β lightweight, zero-dependency store for development and testing
Data Sources
- Snowflake β
SnowflakeIngestorfor table/query ingestion, schema introspection, pagination, and multiple auth methods (password, key-pair, OAuth, SSO) (pip install semantica[db-snowflake])
Document Parsing
- Docling β PDF, DOCX, PPTX, XLSX with table and layout extraction
LLM Providers
- 100+ models via LiteLLM β OpenAI, Anthropic, Cohere, Mistral, Ollama, Azure, AWS Bedrock, and more
AI Frameworks
- Complements LangChain, LlamaIndex, AutoGen, CrewAI, Google ADK
Export
- RDF: Turtle, JSON-LD, N-Triples, XML Β· Parquet Β· ArangoDB AQL
# Core
pip install semantica
# With all optional dependencies
pip install semantica[all]
# Vector store backends (install only what you need)
pip install semantica[vectorstore-pinecone]
pip install semantica[vectorstore-weaviate]
pip install semantica[vectorstore-qdrant]
pip install semantica[vectorstore-milvus]
pip install semantica[vectorstore-pgvector]
# Snowflake ingestion
pip install semantica[db-snowflake]
# From source
git clone https://github.com/Hawksight-AI/semantica.git
cd semantica
pip install -e ".[dev]"
# Run tests
pytest tests/| Channel | Purpose |
|---|---|
| Discord | Real-time help, showcases |
| GitHub Discussions | Q&A, feature requests |
Enterprise support, professional services, and commercial licensing will be available in the future. For now, we offer community support through Discord and GitHub Discussions.
Current Support:
- Community Support - Free support via Discord and GitHub Discussions
- Bug Reports - GitHub Issues
Future Enterprise Offerings:
- Professional support with SLA
- Enterprise licensing
- Custom development services
- Priority feature requests
- Dedicated support channels
Stay tuned for updates!
- AI / ML engineers β GraphRAG, explainable agents, decision tracing
- Data engineers β governed semantic pipelines with full provenance
- Knowledge engineers β ontology management and KG construction at scale
- High-stakes domains β healthcare, finance, legal, cybersecurity, government
All contributions welcome β bug fixes, new features, tests, and docs.
- Fork the repo and create a branch
pip install -e ".[dev]"- Write tests alongside your changes
- Open a PR and tag
@KaifAhmad1for review
See CONTRIBUTING.md for full guidelines.
MIT License Β· Built by Hawksight AI Β· β Star on GitHub