GridGain Developers Hub
GitHub logo GridGain iso GridGain.com
GridGain Software Documentation

Deployment Modes

GridGain clusters and applications can be deployed in just about any environment, whether it’s bare-metal, cloud, Docker, or Kubernetes. GridGain cluster nodes are simply operating system processes that are interconnected over the network to form a single cluster with shared resources such as RAM and CPU. Thus, starting a GridGain cluster is simply a matter of nodes spawning, automatically discovering each other, and forming a shared cluster or storage database.

In this article we discuss primary deployment modes of GridGain clusters: an in-memory data grid (IMDG) and a system of records (aka in-memory database with native persistence). Also, we’ll review two types of deployments related to applications: classical client-server mode and embedded.

However, before discussing the deployment modes let’s review basic concepts such as servers and thick and thin clients. Skip these sections if you are familiar with them.

Servers and Clients

GridGain defines two types of nodes — servers and clients.

A server node is the base computational and data storage unit in GridGain. Typically, you start a single server node per machine or container and it will scale vertically by utilizing all of the CPU, RAM, and other resources available unless specified differently. Those resources are pooled and become available to GridGain applications once the server node joins a cluster of other server nodes.

GridGain Deployment

A cluster is a group of server nodes interconnected together in order to provide shared resources like RAM and CPU to your applications.

Operations executed by applications (key-value queries, SQL, computations, etc.) are directed to and performed by server nodes. If you need more computational power or data storage, scale out your cluster by adding more server nodes to it.

Client nodes or connections are your connection endpoints and gateways from the application layer to the cluster of server nodes. You always embed a client into your application code and execute required APIs. The clients shield all the complexity of GridGain’s distributed nature from application developers who will see the cluster as a single unit. It’s as simple as connecting to an RDBMS via a JDBC driver or Spring Data framework.

Thick vs Thin Clients

GridGain clients come in several different flavors, each with various capabilities. JDBC and ODBC drivers are useful for SQL-only applications and SQL-based tools. Thick and thin clients go beyond SQL capabilities and support many more APIs. Finally, ORM frameworks like Spring Data or Hibernate are also integrated with GridGain and can be used as an access point to your cluster.

Let’s review the difference between thick and thin clients by comparing their capabilities.

Thick clients (aka. standard clients) join the cluster via an internal protocol, receive all of the cluster-wide updates such as topology changes, are aware of data distribution, and can direct a query/operation to a server node that owns a required data set. Plus, thick clients support all of the Ignite and GridGain APIs.

Thin clients (aka. lightweight clients) connect to the cluster via binary protocol with a well-defined message format. This type of client supports a limited set of APIs (presently, key-value and SQL operations only) but in return:

  • Makes it easy to enable programming language support for GridGain and Ignite. Java, .NET, C++, Python, Node.JS, and PHP are supported out of the box.

  • Doesn’t have any dependencies on JVM. For instance, .NET and C++ thick clients have a richer feature set but start and use JVM internally.

  • Requires at least one port opened on the cluster end. Note that more ports need to be opened if partition-awareness is used for a thin client.

Choosing a Client

As a rule of thumb, follow this logic if you are in doubt over which client to use:

  • Java, .NET, and C++ developers - if you need more than key-value and SQL APIs then use the thick client versions which support compute grid, machine learning, transactions, and much more. If key-value and SQL APIs fit your requirements then consider the thin clients instead (note that if partition-awareness is enabled for the thin client then performance should be comparable to the thick clients).

  • JDBC thin vs JDBC thick - use the thin version by default. Fallback to the thick client only if you need faster performance and enabling partition-awareness for the thin driver hasn’t improved performance enough for your use case.

  • Python, Node.JS, PHP, and other programming languages developers - you don’t have any alternatives so your choice is simple: use the existing thin clients.

Thin Client Proxy and Partition Awareness

One of the minor issues with using thin clients is that there can be a performance degradation when applications connect to the cluster without being aware of the data partitioning.

Without partition awareness, an application that is connected to the cluster via a thin client executes all queries and operations via a single proxy server node. These operations must then be re-routed to the correct node. This results in a bottleneck that could prevent the application from scaling linearly.

With partition awareness in place, the same application can directly route queries and operations to primary nodes that own the data required for the queries. This eliminates the bottleneck, allowing the application to scale more easily.

Without Partition Awareness

Notice how queries must pass through the proxy server node, where they are then routed to the correct node.

With Partition Awareness

With partition awareness, queries go directly to the correct node.

Cluster Deployment Modes

GridGain’s unique memory-centric storage architecture provides two primary modes for deployment. The first mode is as an in-memory data grid (IMDG) where GridGain is positioned between your application layer and an external database. This enables in-memory computing for existing solutions in the most straightforward way. The second deployment mode represents GridGain as a classical system of records/hybrid database where GridGain caches data in RAM and stores the records in a native transactional persistence in non-volatile memory or disk.

In-Memory Data Grid

When GridGain is used as a caching layer on top of an external storage, such as a relational database, your are dealing with the in-memory data grid (IMDG) use case. In this case, the whole data set is stored in the external database and what fits in memory is loaded there to boost application performance.

IMDG

As you can see in the figure above, your applications can start treating a GridGain cluster as a primary storage, using the cluster to keep an underlying database in sync. That’s the easiest and least disruptive way to enable in-memory computing for existing solutions - position GridGain between your application and database layers, load data into RAM, and boost performance.

GridGain as an IMDG is different from a simple in-memory cache use case because it allows you to:

  • Migrate your applications faster by accessing your data sets with both standard key-value and SQL APIs.

  • Gain more performance advantages with collocated processing by running any Java-, .NET-, or C++-based logic on the cluster end, avoiding data-movement over the network.

  • Enforce strong consistency whenever is needed - use distributed ACID transactions for your operations and instruct GridGain to keep the underlying storage 100% in sync and consistent.

System of Records

When you use GridGain native persistence instead of an external storage, then you are enabling the in-memory database (IMDB) or system of records use case.

IMDG

In this use case, all of the data is stored on disk and as much as will fit is loaded into RAM. This allows for a much larger data set as data that does not fit in memory is still available. For example, if your data set is large enough that only 10% of it can fit in memory, 100% of the data will be stored on disk and 10% is cached in memory for performance. This configuration, where the data set is stored in bulk on disk, is called GridGain persistence (or native persistence).

This use case provides the following benefits:

  • Multi-tiered storage across RAM and disk: 100% of data is persisted to disk ("warm" and "cold" data sets) while "hot" data always stays in RAM. You lay out the data the way you need. Applications just run queries and GridGain internally goes either to RAM or disk, transparently. This property is crucial for real-time analytics and data warehousing offloading.

  • Instantaneous restarts and advanced high-availability: in the case of full cluster restarts, the cluster becomes fully operational as soon as the nodes are interconnected. No need to preload anything from disk to RAM (based on the advantage above).

Heterogeneous

In some cases you might have a mixed deployment configuration when GridGain, for instance, caches RDBMs tables with read-only access while the tables are updated via the RDBMs directly and all of the changes should be propagated to the cluster. Or, you might have a GridGain cluster that has a subset of the tables that are persisted in an external database (IMDG mode) while the others are persisted to the native persistence.

Usually heterogeneous deployment modes are used for projects when it’s hard to migrate to a clean IMDG or system of records mode and a transitioning or custom architecture is developed.

Application Deployment Modes

When deploying applications that will be interacting with the GridGain cluster, you can continue using a classical client-server mode that separates the lifecycles of applications and databases or you can consider the embedded mode, which is useful for ultra low-latency use cases.

Client-Server Deployment

Client-Server deployment is the classical and most widely used deployment mode, regardless of your deployment environment (bare-metal, VMs, Kubernetes, etc.). This mode separates the lifecycles of your applications and the GridGain cluster. The applications are deployed, developed, updated, and redeployed independently of the cluster of server nodes. The cluster is deployed once and used as a database/storage and computational platform that is restarted only in the case of rare maintenance events.

Overall, the primary benefits of this standard deployment mode are:

  • Separate applications and cluster maintenance lifecycles - as with relational-databases, it lets application developers focus on development and delivery of business logic with different deployment and maintenance rules for the GridGain cluster.

  • Ease of scalability - as long as the cluster is deployed and maintained independently of the applications, it’s easy to scale it out to 10s, 100s, or 1000s of server nodes based on your workloads. The applications connected to the cluster are not impacted while the latter is being scaled.

  • You still can execute custom logic on the cluster end, avoiding data movement over the network by using the compute grid and machine learning APIs.

Embedded Deployment

GridGain clients are always embedded into your application code and run as part of the application process. At the same time it’s possible to embed server nodes into the application process. There are two reasons why you would consider this deployment option.

First, this mode can be used for testing. For instance, you can start a Unit test that will launch several server nodes and carry on with testing scenarios by using this local in-process cluster. The Apache Ignite community and GridGain benefit from this deployment mode for functional testing.

Second, the embedded mode is ideal for ultra low-latency use cases like high-performance mission critical applications or electronic trading. The latency has to stay within the 20-100 microseconds range and any operation that involves network communication can easily go beyond the range. Server nodes embedded into your application instances should store a full copy of the data to ensure all of the application calls are local and not directed to other servers over the network.

So, the pros of the embedded mode are that you can easily achieve microseconds SLAs by avoiding network communication between the applications and the cluster as much as possible. There are, however, some cons to this mode for ultra-low latency scenarios:

  • Limited scalability - as long as each server node needs to store a full copy of the data, it’s impossible to scale the cluster beyond the capacity of a single application instance.

  • Performance implications for write/update operations - even though this mode can keep reads down in the 20-100 microseconds range, the writes/updates still have to be propagated to all of the servers storing the same data replica.

  • Application lifecycle impacts storage lifecycle - any update or restart of the application instances will cause a restart of embedded server nodes (your storage).