Advanced Analytics

Designed for the future.

A simple and powerful analytics tool which lets anyone learn and make decisions from their company’s data.

Answers

Really simple

A simple and powerful analytics tool which lets anyone learn and make decisions from their company’s data—no technical knowledge required.

Self-service analytics for everyone

Point-and-click, drag & drop interface to query data in the most intuitive ways.

Embeddable analytics

Dashboards and visualizations can be embedded with no effort on websites and thrid-party applications.

Features:

Visual Query Builder

Free your data from the confines of SQL, letting everyone query in a UI humans actually love. Need joins, drill-downs and calculated columns? We’ve got you.

Dashboards

Easily create and share interactive dashboards, from TB-scale analytical workloads to day-to-day operational workflows, with 15+ visualization types.

Models

Craft metadata-rich, semantic models which let people query on their own, keeping things consistent and avoiding repetition.

Caching

Go faster with result and model caching. No schedulers or pipelines needed.

Permissions

Control access to data, content and features with group permissions.

Audit

Inspect user behavior across content, data access and data downloads.

Sandboxing

Segment access and unlock multi-tenant analytics via row-level data sandboxes.

Drill-through

Give people an easy and intuitive starting-point to play with data, dig into specifics, and turn it into action.

Auto-analysis

Auto-generated dashboards of questions about your data. Keep what’s interesting, hedge what’s not, add more questions, and text boxes.

Explorer

A data store for sparse connected data​ built for AI

The shortest path to insights

Fast, interactive data exploration capabilities to quickly find hidden connections and insights in your data.

Unified visual interface

BigConnect Explorer provides tools for data discovery and analysis, collaboration, alerting and information management.

Extensible

Can be extended to suit any use case, using common development languages like Java, JavaScript, React, CSS/LESS etc. Customizations are in the form of plugins that can be developed both for the back-end and the front-end.

Features:

Dynamic Data Model

Powered by a dynamic, semantic data model to store, correlate and query all information. It’s a conceptual data model that expresses information in a factual way: objects, relations and attributes.

Workspaces

Make your work private with spaces. Choose who you invite and work together on a specific use-case. Everything you do is confined to the selected workspace.

Collaboration

Invite people to work together, see their changes in realtime, add comments and share knowledge.

Analytics

Use the available tools to make sense of unstructured data: link analysis, spatial analysis, charts, timeline and powerful search.

Enterprise Search

Google-like search, multi-faceted filtering, visual query builder that works with spatial queries and much more to find anything really fast.

Load data

Connect to databases or drag & drop files to your workspace. Smart mapping wizards let you map data immediately after you load it.

Enrichment

Powerful text and image enrichment capabilities to annotate your information, creating new concepts, properties and relations.

Administration

The administration interface allows you to customize the data model, security, enrichment and much more.

Extensibility

Designed to be extensible with plugins, frontend and backend. A lot of hooks and extension points are available to customize it to your specific needs.

Pipelines

Data Integration for Hybrid & Multi-Cloud Environments

Data Ingestion Pipelines, Simplified

Spend more time building data smart pipelines, enabling self-service and innovating without the noise. BigConnect Pipelines is an easy-to-use data pipeline engine for streaming, CDC and batch ingestion from any source to any destination.

  • Build pipelines for streaming, batch and change data capture (CDC) in minutes
  • Eliminate 90% of break-fix and maintenance time
  • Port data pipelines to new data platforms without rewrites

Operationalize Your Data Collection

Single Experience for All Design Patterns

Build schema-agnostic smart data pipelines with pre-built sources and destinations in minutes for streaming, batch, and change data capture (CDC), using a single, visual tool.

Ingest Data Across Multiple Platforms

Run your data in a development environment on multiple platforms without rework. Pipelines are platform agnostic by design so you can reuse them across data platforms in hybrid and multi-cloud environments.

Built for Change

Worst case scenario: an upstream change doesn’t break your pipeline, it flows unreliable, incorrect, or unusable data into your analytics platform undetected. Intent-driven pipelines built for data drift, reducing risk of bad data downstream and outages. When data drift happens, pipelines alert you to remediate issues or embrace emergent design.

Typical use cases:

Load data into data warehouses

Import and prepare data for analysis in data warehouses with automatic transformation capabilities.

Ingest data into data lakes

Ingest unstructured and object data into data lakes from many sources

Integrate with big data for machine learning (ML) and analytics

Import and prepare data for use in modeling and big-data analytics with Hadoop, Kafka, JMS, NoSQL databases for ML and AI.

Modernize data infrastructure

Enjoy native support for big data and AI by leveraging the BigConnect platform for integrations.

Features:

High-performance data transformation

Powerful ETL

Boxed with features to build, monitor and debug your data pipelines.

Distributed

Run pipelines in a clustered environment to accelerate your ETL flows

Native big data support

Transform and execute data mappings.with native BigConnect features for super-fast data transfer.

Features:

Out-of-the-box integrations

Prebuilt connectors

Large library of technologies simplifies integrating sources and targets to give developers more flexibility.

Declarative design

Create data-mapping rules to accelerate development of integration logic and increase productivity.

Guided integration

Best practices to simplify data mapping between systems

Features:

Heterogeneous system support

Big data integration

Prebuilt big data connectors for Hadoop, Spark, Hive, Kafka, HBase, Sqoop, Pig, Cassandra, NoSQL databases, and more allow users to integrate any data from any source.

Popular databases and other technologies

Extensive database support, ERP and CRM systems, B2B systems, flat files, XML, JSON, LDAP, JDBC, ODBC, SaaS, and third-party appliances.

Enterprise Crawling

Data Integration for Hybrid & Multi-Cloud Environments

Turn the web into data

Whether is about posts, web sites, blogs or forums, Sponge is a full-featured, flexible and extensible visual web crawler that will help you gather information from web sites and social media.

High Performance

Using multiple crawling threads and distributed clusters, incremental crawling and multithreading, data can be quickly scraped from thousand of web sites.

Smart data collection

Simulare real user interactions, bypassing IP blocks, CAPTCHAs, geo-restrictions, anti-bots, and more. Perfect for seamless public data collection or multi-account management.

Powered by AI

Don't worry about manually annotating website layout to extract what you need. Our AI models will understand any website and automatically extract what you need.

Popular use cases

Turn websites into APIs

Need to regularly extract data from a website or web app that doesn’t have an API? Create your own scalable API with no code.

Price monitoring

Track competitor pricing, discounts, and promotions to stay ahead in the market and make data-driven pricing decisions.

Product data extraction

Collect product information such as names, descriptions, images, and specifications from e-commerce websites, making it easier to compare products or perform market analysis.

Lead generation

Extract contact information from business directories to generate a list of potential leads for sales and marketing purposes.

Sentiment analysis

Extract reviews, comments, or social media posts to analyze customer sentiment and feedback, enabling businesses to make informed decisions and improve their offerings.

News and content aggregation

Monitor news websites, blogs, or forums to stay updated on industry trends, collect research data, or find content for curation.

Social media monitoring

Monitor social media platforms for mentions of your brand, products, or competitors to stay informed about customer opinions and market trends.

Core

A data store for sparse connected data​ built for AI

99.99%

High Availability

Trillions

Nodes and Edges

Millions

Queries Per Second

Millisecond

Latency

Multi-modal data store

Store any type of structured, semi-structured and unstructured data. From database tables to documents, images, audio and video, and raw binary data.

Multi-modal query and search

With native support for full-text search, Cypher, SQL and GQL you can query the data store in any possible way. And really, really fast.

Secure by design

Apply security labels to attributes, nodes and edges and combine them using logical operators to implement any imaginable scenario. Combine this with storage-level encryption and secure communications and you will have the most secure data store in the world.

Horizontally scalable

Built on a shared-nothing distributed architecture to offer linear scalability, meaning that you can add more nodes or services to the cluster without affecting performance. Our native engine enables lightning-fast QPS and TPS with millisecond latency. With high concurrency, efficient traversals, and optimized memory usage, you can process sheer volumes of data blazingly fast.

Resilient to failures

With horizontal scalability and snapshots feature, it guarantees high availability even in case of failures. The snapshot feature allows users to take snapshots at any point in time so that they can use them to restore data at that specific time. If a disaster occurs, it can be recovered from snapshots without data loss. This process is fully automatic and transparent to clients.

Typical Use Cases:

LLM Semantic Memory

Provides long-term memory to Large Language Models in the form semantic knowledge graphs.

Explainable AI

One of the biggest obstacles to widespread artificial intelligence adoption is a lack of transparency as to how the AI system arrived at a particular decision. Using LLM memory, our platform can track how Generative AI perform Chain of Thought reasoning to reach a conclusion.

Customer 360

Enables marketing teams to connect different types of information to get a complete overview of how customers interact with their systems.

Anti-Money Laundering

Money launderers typically create an intricate network of identities and accounts to funnel their ill-gotten gains. Deep Link Analytics helps identify fraud rings, suspicious transactions and unusual behavior.

Cybersecurity Threat Detection

Any network is a network of components and processes: the internet is an interconnected system of servers, routers, bridges, laptops, smartphones, and so on – and there are processes defining how these work together. Deep Liks Analytics helps identifying patterns of behavior associated with malicious attacks.

Risk Assessment & Monitoring

Assess and monitor Credit Risk and Regulatory Risk for customers and suppliers

Energy Management

From meter readings to information from network sensors, balancing a power grid requires consolidating signals from multiple levels of the power infrastructure and matching demand and supply with complex linear equations. Network operators can respond immediately to sudden spikes in demand or drops in supply, thus reducing operational risk and operating costs while improving reliability, efficiency, and customer experience.

Network & IT Management

Detect when a resource such as a storage array, server, network switch or router shows the signs of wear, requires maintenance or is nearing its peak capacity.

Supply Chain Analysis

Supply and delivery pipelines have dozens, if not hundreds, of stages and an ability to analyze and understand the impact across many levels is essential. Our platform solution has advanced analysis and pattern recognition to identify product delays, shipment status, and other quality control and risk issues.

Features:

Symmetrically distributed

Data is symetrically distributed accross storage nodes based on optimal shards and partitions and Multi Group Raft architecture that take maximum advantage of NVMe SSDs

Storage and computing separation

Storage nodes are separated from compute nodes to ensure maximum load distribution in a multi-node deployment configuration.

Horizontal scalability

With a shared-nothing distributed architecture, NebulaGraph offers linear scalability, meaning that you can add more nodes or services to the cluster without affecting performance.

Cypher and SQL

Support for both OpenCypher and ANSI SQL allows developers to integrate BigConnect into any application.

Full-text search

Full-text indexes are powered by Elasticsearch. This means you can use Elasticsearch the full-text query language to retrieve what you want.

GQL query language

A declarative graph query language designed for both developers and operations, that allows expressive and efficient graph patterns.

Distributed Analytics

Based on Apache Spark, allows massively parallel graph computations and the full benefits of Spark's data processing, transformation and machine learning pipelines.

Bakcup & Snapshots

One-click backup and restore to local filesystem or S3-compatible cloud providers. Support for snapshots allows poin-in-time recovery.

Asynchronous Processing

An event-based data processing runtmie allows late processing of data, while preservind order of operations. Very useful for Machine Learning and AI data augmentation.

Gaph embeddings

Low-dimensional, compact graph representations that store relational and structural data in a vector space, condensing complicated graph structures into dense vectors.

Metrics

Query database metrics over HTTP to get a comprehensive overview of performance and availability of the entire data store.

Security

Attribute, node and edge security labels can be applied and combined using expressions to achieve maximum flexibility. Supports also storage encryption and SSL comunication.