OrientDB Manual - version 3.2.33
Welcome to OrientDB - the first Multi-Model Open Source NoSQL DBMS that brings together the power of graphs and the flexibility of documents into one scalable high-performance operational database.
OrientDB v 3.2.33 is our latest GA
Quick Navigation
Operations
- Installation
- 3rd party Plugins
- Upgrade
- Configuration
- Distributed Architecture (replication, sharding and high-availability)
- Performance Tuning
- ETL to Import any kind of data into OrientDB
- Import from Relational DB
- Backup and Restore
- Export and Import
Quick References
- Console
- Studio web tool
- OrientDB Server
- Network-Binary-Protocol
- Gephi Graph Analysis Visual tool
- Rexster Support and configuration
- Continuous integration
Resources
- User Group - Have question, troubles, problems?
- #orientdb IRC channel on freenode
- Professional Support
- Training - Training and classes.
- Events - Follow OrientDB at the next event!
- Team - Meet the team behind OrientDB
- Contribute - Contribute to the project.
- Who is using OrientDB? - Clients using OrientDB in production.
Questions or Need Help?
Check out our Get in Touch page for different ways of getting in touch with us.
Past Releases
Every effort has been made to ensure the accuracy of this manual. However, OrientDB, LTD. makes no warranties with respect to this documentation and disclaims any implied warranties of merchantability and fitness for a particular purpose. The information in this document is subject to change without notice.
Introduction
This Chapter provides an introduction to OrientDB.
- About This Manual
- Overview of OrientDB
- Release 3.2
- How to Report Bugs or Problems
- How to Get in Touch with OrientDB
- How to Contribute to OrientDB
- OrientDB Roadmap
About this Manual
This Manual documents version 3.2 of the OrientDB Server and the OrientDB suite of products.
Overview
OrientDB is the first Multi-Model Open Source NoSQL DBMS that combines the power of graphs and the flexibility of documents into one scalable, high-performance operational database.
Gone are the days where your database only supports a single data model. As a direct response to polyglot persistence, multi-model databases acknowledge the need for multiple data models, combining them to reduce operational complexity and maintain data consistency. Though graph databases have grown in popularity, most NoSQL products are still used to provide scalability to applications sitting on a relational DBMS. Advanced 2nd generation NoSQL products like OrientDB are the future: providing more functionality and flexibility, while being powerful enough to replace your operational DBMS.
Speed
OrientDB was engineered from the ground up with performance as a key specification. It’s fast on both read and write operations. Stores up to 120,000 records per second
- No more Joins: relationships are physical links to the records.
- Better RAM use.
- Traverses parts of or entire trees and graphs of records in milliseconds.
- Traversing speed is not affected by the database size.
Enterprise
While most NoSQL DBMSs are used as secondary databases, OrientDB is powerful and flexible enough to be used as an operational DBMS. OrientDB Enterprise Edition gives you all the features of our community edition plus:
- Incremental backups
- Unmatched security
- Query Profiler
- Distributed Clustering configuration
- Metrics Recording
- Live Monitor with configurable alerts
Zero Configuration Multi-Master Architecture
With a master-slave architecture, the master often becomes the bottleneck. With OrientDB, throughput is not limited by a single server. Global throughput is the sum of the throughput of all the servers.
- Multi-Master
- Elastic Linear Scalability
- Restore the database content using WAL
With a zero-config multi-master architecture, OrientDB is perfect for the Cloud. Hundreds of servers can share the workload, scaling horizontally across distributed modern data centers.
Flexibility
Replacing your DBMS, once it no longer meets requirements, can be a huge cost in time and resources. Is your database powerful, scalable and flexible enough to grow with you? Using OrientDB also reduces the need to support multiple products to achieve your goals.
Get Started in Minutes
OrientDB is written entirely in Java and can run on any platform without configuration and installation. It's a drop-in replacement for the most common existing graph databases in deployment today.
- Multiple Programming Language Bindings
- Extended SQL with graph functionality
- Tinkerpop API
Coming from the relational world? Import your data using OrientDB Teleporter and get started in an instant
Coming From Neo4j? Our Neo4j Importer lets you get started with your own graph data
Low TCO
There is absolutely no cost associated with using OrientDB Community Edition.
- OrientDB Community is free for commercial use.
- Comes with an Apache 2 Open Source License.
- Eliminates the need for multiple products and multiple licenses.
OrientDB Ltd, the company behind OrientDB, offers optional services such as Developer and Production Support, consultancy and Training with transparent pricing far below our competitors to ensure you’re maximizing OrientDB’s capabilities for your particular use case. OrientDB Enterprise Edition is included with the purchase of any of these services.
Open Source
Which is more likely to have better quality? A DBMS created and tested by a handful of developers or one tested by over 100,000 developers globally? When code is public, everyone can scrutinize, test, report and resolve issues. All things Open Source move faster compared to the proprietary world. In fact, the most popular databases ranked by DB-Engines are now Open Source.
Editions
OrientDB is available in two editions:
-
Community Edition is released as an open source project under the Apache 2 license. This license allows unrestricted free usage for both open source and commercial projects.
-
Enterprise Edition used to be the commercial software built on top of the Community Edition, now it is Free Open Source as well. Enterprise is developed by the same team that developed the OrientDB engine. It serves as an extension of the Community Edition, providing Enterprise features, such as:
- Non-Stop Backup and Restore
- Scheduled FULL and Incremental Backups
- Query Profiler
- Distributed Clustering configuration
- Metrics Recording
- Live Monitoring with configurable Alerts
Release 3.2
This Section documents what is new and what has changed in OrientDB 3.2.
- What's new in OrientDB 3.2?
- Available Packages
- SQL Changes
- API Changes
- Known Issues
- Upgrading to OrientDB 3.2
What's new in OrientDB 3.2?
Security
With the OrientDB 3.2 multimodel API some important changes were implemented on Database creation. The most important one is that basic database users (admin, reader, writer) are no longer created by default. This is the consequence of some considerations about Security: having default passwords (eg. admin/admin) is a bad practice and exposes OrientDB installations to possible attacks.
Of course it's possible to explicitly trigger user creation and to provide a custom password, instead of using a default admin user.
final OrientDB orientdb = ...;
orientdb.execute("create database test plocal users ( admin identified by 'adminpwd' role admin)");
final ODatabaseSession session = orientdb.open("test","admin", "adminpwd");
the roles admin
, writer
, reader
are still created by default.
It is highly recommended to take a stronger password than
adminpwd
in our example.
If the creation of default users is enabled and you try to create a user called admin
, the creation of that user will fail.
The creation of default users can be disabled setting CREATE_DEFAULT_USERS
to false
as in:
new OrientDB("...",
OrientDBConfig.builder()
.addConfig(OGlobalConfiguration.CREATE_DEFAULT_USERS, false)
.build());
The creation of multiple admin users like admin
, reader
, and writer
can be done by comma-separating
final OrientDB orientdb = ...;
orientdb.execute("create database test plocal users (
admin identified by 'adminpwd' role admin,
reader identified by by 'adminpwd' role reader,
writer identified by by 'adminpwd' role writer)");
final ODatabaseSession session = orientdb.open("test","admin", "adminpwd");
From Studio, there is a new option that allows you to define the default admin password on the Database Create dialog.
The old defaults can be restored (for backward compatibility) by setting -Dsecurity.createDefaultUsers=true
at startup
Deprecated APIs like
ODatabaseDocumentTx
are not affected from that change.
Server-Level Commands
In OrientDB v 3.2 we added some infrastructure to run commands at server level. See SQL Changes and Server-Level Commands)
Distributed enhancements and stabilization
In v 3.1.x, OrientDB clustering had a limitation related to indexing: UNIQUE indexes needed full agreement between the nodes (N ack out of N nodes) to perform a transaction. In v 3.2 we removed this limitation, now transactions that involve UNIQUE indexes have the same quorum as any other transaction (typically N/2 + 1). Hazelcast was updated to latest version.
Console enhancements
In previous versions there are some slight differences in the console when using a remote server or an embedded environment. In v 3.2 the CLI interface was refactored to unify the management of embedded and remote environments
see
GraalVM support
After Nashorn was deprecated and then removed from the JVM (see JSR 335), OrientDB needs an alternative for scripting. In v 3.2 we introduce support for GraalVM as a replacement for scripting capabilities
Available Packages
Starting from OrientDB 3.0, for each OrientDB Editions, different packages are available for download. Please use the package that best fits your needs.
Community Edition
The following are the packages available for OrientDB Community Edition:
- CE (orientdb-community-3.2.33.tar.gz)
- OrientDB Community Edition without Gremlin/TinkerPop support
- CE-TP2 (orientdb-community-tp2-3.2.33.tar.gz)
- OrientDB Community Edition that includes support for Gremlin/TinkerPop 2.6
- CE-TP3 (orientdb-community-tp3-3.2.33.tar.gz)
- OrientDB Community Edition that includes support for Gremlin/TinkerPop 3.x
Enterprise Edition
- Agent (agent-3.2.33.tar.gz)
- OrientDB Enterprise Agent
SQL Changes
Server-level Commands
In previous versions, you can run scripts (eg. SQL) on a single database connection, and the scripts are executed on the database itself. In OrientDB v 3.2 we extended the usage of scripts to the Server level.
This adds a lot of new potential for both infrastructure management and querying.
For now, we implemented some basic Server-level commands, like CREATE/DROP DATABASE
and CREATE SYSTEM USER
, but the infrastructure allows for potential future extensions to all the aspects of the server and data management.
Server-level commands can be executed via native API, via REST or via CLI
API Changes
Known Issues
Upgrading to OrientDB 3.2
Binary formats
When upgrading OrientDB to a newer major/minor version, we recommend to perform an export and import of the databases. Anyway, OrientDB guarantees binary format compatibility, so export/import is not striclty needed.
General information
Developing OrientDB 3.2 we put a lot of attention on maintaining backward compatibility with v 3.0 and 3.1.
Here is a list of the things you should know when migrating to v 3.2
Database Creation
In OrientDB v 3.2, the creation of a new database does not automatically provide default users (admin, reader, writer). This choice is due to Security considerations: default users come with default passwords, that are a possible weakness in the server security if not promptly changed.
A new database API is provided to create custom users at DB creation time:
Orientdb orientdb = ...;
orientdb.execute("CREATE DATABASE foo plocal users(admin identified by 'adminpwd' role admin")
The old behaviour (ie. creating default users with default password) can be restored by setting the Global Configuraion option called security.createDefaultUsers
:
eg.
./server.sh -Dsecurity.createDefaultUsers=true
Console
The console was adapted to include support to Server-Level commands and to allow the creation of databases without default users (see above).
The old create database
command had a complex behaviour: it created a DB with default users and then connected to that database using admin
user and the default password. This is not possible anymore by default (default admin
user does not exist anymore), so create database
no longer connects to the DB.
The new console interaction pattern involves connecting to the server/environment (see CONNECT ENV) and then execute server-level commands on it.
CREATE DATABASE
command was enhanced to accept default user names and password (see CREATE DATABASE)
A new OPEN <database>
command was added to connect to an existing database in current server/environment (see OPEN)
A backward compatibility option is provided to allow execution of old console scripts; it can be enabled setting the console configuration as follows (this can just be added as the first row of the script):
orientdb> SET compatibilityLevel=0;
Release notes
General information on how to upgrade OrientDB can be found in the Upgrade Chapter.
You may also be interested in checking the Release Notes.
Report an Issue
Very often when a new issue is open it lacks some fundamental information. This slows down the entire process because the first question from the OrientDB team is always "What release of OrientDB are you using?" and every time a Ferret dies in the world.
So please add more information about your issue:
- OrientDB release? (If you're using a SNAPSHOT please attach also the build number found in "build.number" file)
- What steps will reproduce the problem?
- Settings. If you're using custom settings please provide them below (to dump all the settings run the application using the JVM argument -Denvironment.dumpCfgAtStartup=true)
- What is the expected behavior or output? What do you get or see instead?
- If you're describing a performance or memory problem the profiler dump can be very useful (to dump it run the application using the JVM arguments -Dprofiler.autoDump.reset=true -Dprofiler.autoDump.interval=10 -Dprofiler.enabled=true)
Now you're ready to create a new one: https://github.com/orientechnologies/orientdb/issues/new
How to Get in Touch with OrientDB
We want to make it super-easy for OrientDB users and contributors to talk to us and connect with each other, to share ideas, solve problems and help make OrientDB awesome. Here are the main channels we're running currently, we'd love to hear from you on one of them:
Discussions
The Discussions Group (aka Community Group) is a good first stop for a general inquiry about OrientDB or a specific support issue (e.g. trouble setting OrientDB up). It's also a good forum for discussions about the roadmap or potential new functionality.
StackOverflow
Feel free to ask your questions on StackOverflow under "orientdb" and "orient-db" tags.
Follow and chat to us on Twitter.
GitHub
If you spot a bug, then please raise an issue in our main GitHub project orientechnologies/orientdb. Likewise if you have developed a cool new feature or improvement in your OrientDB fork, then send us a pull request against the "develop" branch!
If you want to brainstorm a potential new feature, then the OrientDB Discussions Group (see above) is probably a better place to start.
How to Contribute to OrientDB
In order to contribute issues and pull requests, please sign OrientDB's Contributor License Agreement. The purpose of this agreement is to protect users of this codebase by ensuring that all code is free to use under the stipulations of the Apache2 license.
Pushing into main repository
OrientDB uses different branches to support the development and release process.
The develop
branch contains code under development for which there's not a stable release yet.
When a stable version is released, a branch for the hotfix is created.
Each stable release is merged on master branch and tagged there.
At the time of writing these notes, the state of branches is:
- develop: work in progress for next 2.2.x release (2.2.0-SNAPSHOT)
- 2.1.x: hot fix for next 2.1.x stable release (2.1.10-SNAPSHOT)
- 2.0.x: hot fix for next 2.0.x stable release (2.0.17-SNAPSHOT)
- last tag on master is 2.1.9
If you'd like to contribute to OrientDB with a patch follow the following steps:
- fork the repository interested in your change. The main one is https://github.com/orientechnologies/orientdb, while some other components reside in other projects under Orient Technologies umbrella.
- clone the forked repository
- select the branch, e.g the develop branch:
git checkout develop
- apply your changes with your favourite editor or IDE
- test that Test Suite hasn't been broken by running:
mvn clean test
- if all the tests pass, then do a Pull Request (PR) against the branch (e.g.: "develop") on GitHub repository and write a comment about the change. Please don't send PR to "master" because we use that branch only for releasing
Documentation
If you want to contribute to the OrientDB documentation, the right repository is: https://github.com/orientechnologies/orientdb-docs. Every 24-48 hours all the contributions are reviewed and published on the public documentation.
Code formatting
v 3.1 and following
Since v 3.1, OrientDB uses Google code formatter.
In IntelliJ Idea, you can use this plugin https://plugins.jetbrains.com/plugin/8527-google-java-format
From Maven, you can run mvn com.coveo:fmt-maven-plugin:format
for automatic code format.
v 3.0 and previous releases
For previous versions (until 3.0) you can use eclipse java formatter config file, that you can find here: _base/ide/eclipse-formatter.xml.
If you use IntelliJ IDEA you can install this plugin and use formatter profile mentioned above.
Debugging
Run OrientDB as standalone server
The settings to run OrientDB Server as stand-alone (where the OrientDB's home is /repositories/orientdb/releases/orientdb-community-2.2-SNAPSHOT
) are:
Main Class: com.orientechnologies.orient.server.OServerMain
VM parameters:
-server
-DORIENTDB_HOME=/repositories/orientdb/releases/orientdb-community-2.2-SNAPSHOT
-Dorientdb.www.path=src/site
-Djava.util.logging.config.file=${ORIENTDB_HOME}/config/orientdb-server-log.properties
-Dorientdb.config.file=${ORIENTDB_HOME}/config/orientdb-server-config.xml
-Drhino.opt.level=9
Use classpath of module: orientdb-graphdb
Run OrientDB distributed
The settings to run OrientDB Server as distributed (where the OrientDB's home is /repositories/orientdb/releases/orientdb-community-2.2-SNAPSHOT
) are:
Main Class: com.orientechnologies.orient.server.OServerMain
VM parameters:
-server
-DORIENTDB_HOME=/repositories/orientdb/releases/orientdb-community-2.2-SNAPSHOT
-Dorientdb.www.path=src/site
-Djava.util.logging.config.file=${ORIENTDB_HOME}/config/orientdb-server-log.properties
-Dorientdb.config.file=${ORIENTDB_HOME}/config/orientdb-server-config.xml
-Drhino.opt.level=9
-Ddistributed=true
Use classpath of module: orientdb-distributed
In order to debug OrientDB in distributed mode, changed the scope to "runtime" in file distributed/pom.xml:
<groupId>com.orientechnologies</groupId>
<artifactId>orientdb-graphdb</artifactId>
<version>${project.version}</version>
<scope>runtime</scope>
In this way IDE like IntelliJ can start the server correctly that requires graphdb dependency.
OrientDB Roadmap
This page contains the roadmap with the main enhancements for the OrientDB product.
Terms
- RC: Release Candidate, is a beta version with potential to be a final product, which is ready to release unless significant bugs emerge. In this stage of product stabilization, all product features have been designed, coded and tested through one or more beta cycles with no known showstopper-class bug. A release is called code complete when the development team agrees that no entirely new source code will be added to this release. There could still be source code changes to fix defects, changes to documentation and data files, and peripheral code for test cases or utilities. Beta testers, if privately selected, will often be credited for using the release candidate as though it were a finished product. Beta testing is conducted in a client's or customer's location and to test the software from a user's perspective.
- GA: General Availability, is the stage where the software has "gone live" for usage in production. Users in production are suggested to plan a migration for the current GA evaluating pros and cons of the upgrade.
Release 3.0
- Development started on.: June 2016
- Expected first M1......: January 2017
- Expected first M2......: February 2017
- Expected first RC......: March 2017
- Expected final GA......: March/April 2017
Status
Last update: December 14, 2016
For a more detailed an updated view, look at the Roadmap 3.0 issue.
Module | Feature | Status |
---|---|---|
Core | Multi-Threads WAL | 30% |
Core | WAL Compaction | 30% |
Core | Index rebuild avoid using WAL | 0% |
Core | Compression of used space on serialization | 3% |
Core | Improved DISKCACHE algorithm | 60% |
Core | Index per cluster | 0% |
Core | New data structure to manage edges | 0% |
SQL | Distributed SQL Executor | 70% |
SQL | Multi-line queries in batch scripts | 100% |
Java API | New factories | 100% |
Java API | Improve SQL UPDATE syntax | 100% |
Java API | Support for TinkerPop 3 | 70% |
Remote protocol | Support for server-side transactions | 10% |
Remote protocol | Support for server-side cursors | 90% |
Remote protocol | Push messages on schema change | 0% |
Remote protocol | Push messages on record change | 0% |
Distributed | Auto-Sharding | 10% |
Distributed | Optimized network protocol to send only the delta between updates | 50% |
Release 3.1
- Development started on.: -
- Expected first RC......: TBD
- Expected final GA......: TBD
Status
Last update: April 12, 2015
Module | Feature | Status |
---|---|---|
Core | Parallel Transactions | 0% |
Core | Indexing of embedded properties | 0% |
Core | Override of properties | 0% |
Core | Enhance isolation level also for remote commands | 0% |
Distributed | Optimized replication for cross Data Center | 0% |
Distributed | Replication of in-memory databases | 0% |
Lucene | Faceted search | 20% |
Java API | ODocument.update() | 0% |
SQL | shortestPaths() function | 0% |
SQL | New functions (strings, maths) | 40% |
OrientDB in Five Minutes
Learn OrientDB basics in just five minutes, in your favorite programming language:
OrientDB for Java Developers in Five Minutes
If you are a Java Developer and it's the first time you approach OrientDB, then you are in the right place!
Ready? Let's start!
Prerequisites
To run OrientDB you will need a Java SDK installed on your machine. OrientDB runs with Java from Version 8 to 11.
Step 1/5 - Download and Install
Download OrientDB from the following URL:
Unzip it on your FileSystem and open a shell in the directory.
Now type
cd orientdb-3.2.0
cd bin
(change the first cd
with the exact version you downloaded)
and then, if you are on Linux/OSX, you can start the server with
./server.sh
if you are on Windows, start the server with
server.bat
You will see OrientDB starting
.` `
, `:.
`,` ,:`
.,. :,,
.,, ,,,
. .,.::::: ` ` ::::::::: :::::::::
,` .::,,,,::.,,,,,,`;; .: :::::::::: ::: :::
`,. ::,,,,,,,:.,,.` ` .: ::: ::: ::: :::
,,:,:,,,,,,,,::. ` ` `` .: ::: ::: ::: :::
,,:.,,,,,,,,,: `::, ,, ::,::` : :,::` :::: ::: ::: ::: :::
,:,,,,,,,,,,::,: ,, :. : :: : .: ::: ::: :::::::
:,,,,,,,,,,:,:: ,, : : : : .: ::: ::: :::::::::
` :,,,,,,,,,,:,::, ,, .:::::::: : : .: ::: ::: ::: :::
`,...,,:,,,,,,,,,: .:,. ,, ,, : : .: ::: ::: ::: :::
.,,,,::,,,,,,,: `: , ,, : ` : : .: ::: ::: ::: :::
...,::,,,,::.. `: .,, :, : : : .: ::::::::::: ::: :::
,::::,,,. `: ,, ::::: : : .: ::::::::: ::::::::::
,,:` `,,.
,,, .,`
,,. `, GRAPH DATABASE
`` `.
`` orientdb.com
`
2017-08-14 14:11:12:824 INFO Loading configuration from: /Users/luigidellaquila/temp/orient/orientdb-community-3.0.0m2/config/orientdb-server-config.xml... [OServerConfigurationLoaderXml]
2017-08-14 14:11:12:932 INFO OrientDB Server v3.0.0 (build 4abea780acc12595bad8cbdcc61ff96980725c3b) is starting up... [OServer]
2017-08-14 14:11:12:951 INFO OrientDB auto-config DISKCACHE=12.373MB (heap=1.963MB direct=524.288MB os=16.384MB) [orientechnologies]
2017-08-14 14:11:12:994 INFO Databases directory: /Users/luigidellaquila/temp/orient/orientdb-community-3.0.0m2/databases [OServer]
2017-08-14 14:11:13:017 INFO Creating the system database 'OSystem' for current server [OSystemDatabase]
2017-08-14 14:11:14:457 INFO Listening binary connections on 0.0.0.0:2424 (protocol v.37, socket=default) [OServerNetworkListener]
2017-08-14 14:11:14:459 INFO Listening http connections on 0.0.0.0:2480 (protocol v.10, socket=default) [OServerNetworkListener]
+---------------------------------------------------------------+
| WARNING: FIRST RUN CONFIGURATION |
+---------------------------------------------------------------+
| This is the first time the server is running. Please type a |
| password of your choice for the 'root' user or leave it blank |
| to auto-generate it. |
| |
| To avoid this message set the environment variable or JVM |
| setting ORIENTDB_ROOT_PASSWORD to the root password to use. |
+---------------------------------------------------------------+
Root password [BLANK=auto generate it]: *
The first time you start the server, you will be asked to enter a root password (twice). You can choose the password you prefer, just make sure to remember it, you will need it alter.
Now you are ready for the Next Step - Create a Database >>>
OrientDB for Java Developers in Five Minutes
In the Previous Step you installed and started OrientDB on your machine, now it's time to create a database
Step 2/5 - Create a DB
Open your favourite browser (modern browser, no IE please!) and enter the following URL:
http://localhost:2480/studio/index.html
You will see this page:
- Click on "NEW DB"
- enter "test" as database name
- enter the root password
- Then click on "CREATE DATABASE"
IMPORTANT: In v 3.2 new databases are created without default db users (admin/reader/writer) to avoid possible vulnerabilities due to default passwords. In this dialog you can select "Create Admin user" and enter a custom admin password for this user.
Good job! You just created your first database!
Now, to make sure that everything is OK, try to enter the following query
SELECT * FROM OUser
and click the green "RUN" button. You should see the three default users that OrientDB creates for you: admin, reader and writer
Now you are ready for the Next Step - Create your Java Application >>>
OrientDB for Java Developers in Five Minutes
In the previous step you created your first database, now it's time to create a Java application that can connect to the DB, query and manipulate it.
From now on, for this tutorial we will use IntelliJ Idea, but you can use any IDE that supports Java 8 and Maven.
Step 3/5 - Create a Maven project
Open IntelliJ Idea and:
- choose the menu File -> new -> Project...
- from the left menu, choose Maven
- make sure that Project SDK is greater or equal to 1.8
- click Next
- choose your GroupId and ArtifactId (we will use "test" "test", as this is what we are doing ;-) )
- click Next twice
Now open pom.xml file and add the following dependency:
<dependencies>
<dependency>
<groupId>com.orientechnologies</groupId>
<artifactId>orientdb-client</artifactId>
<version>3.0.0</version>
</dependency>
</dependencies>
IMPORTANT: make sure you are using the same client version as the server you are using.
and enable Java 8 as target compiler
<build>
<plugins>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId>
<configuration>
<source>1.8</source>
<target>1.8</target>
</configuration>
</plugin>
</plugins>
</build>
Click on "automatically import dependencies" if needed
Now you are ready for the Next Step - Open a DB Connection and Create the DB Schema >>>
OrientDB for Java Developers in Five Minutes
In the previous step you created your Java application scaffolding in IntelliJ Idea, now it's time to create a Java class that connects to the database and does basic operations
We will write a very simple application with two classes:
- Person: a vertex type containing information about people
- FriendOf: an edge class that connects people together
Step 4/5 - Connect to the DB and create a basic schema
Let's create a simple Java class
- Open the folder src -> main -> java
- Right-click on the java folder and choose New -> Java Class
- Give the class the name you prefer and click OK
- create a main method
Connecting to the DB
Now it's time to connect to OrientDB
- Create an OrientDB object to manage the remote server
- Also invoke a
close()
at the end of the method (so that you don't forget it later)
OrientDB orient = new OrientDB("remote:localhost", OrientDBConfig.defaultConfig());
//let's do something with this server!
orient.close();
The first argument of OrientDB()
constructor is the URL of the remote database. It's made of two parts:
remote
: specifies that we are connecting to a remote (stand-alone) serverlocalhost
: the address of the host where OrientDB is running (it can be a URL or an IP, in this case orientdb is running on the local machine)
Until now, we just established a connection with the server, but not with the database itself (we created a db named test, do you remember?).
Now it's time to open a database session (and remember to close it at the end!):
OrientDB orient = new OrientDB("remote:localhost", OrientDBConfig.defaultConfig());
ODatabaseSession db = orient.open("test", "admin", "admin");
//let's do something with this session!
db.close();
orient.close();
Here we are actually connecting to the database. The three parameters are:
test
: the database name (we created it a couple of steps before)admin
(the first one): the username used to connectadmin
(the second one): the password for the connection
By default, when you create a new database, OrientDB creates three users for you: admin
, reader
, writer
; the passwords
for these users are the same as the user names, eg. the password for admin
is admin
. You can change it later of course, and you can
define more users if needed.
Creating the schema
Now it's time to create a schema for our db. Do you remember? We are creating a graph of people who are friends, so we need a Person
vertex class and a FriendOf
edge class.
Let's also add a little check to make sure that the classes do not already exist
if (db.getClass("Person") == null) {
db.createVertexClass("Person");
}
if (db.getClass("FriendOf") == null) {
db.createEdgeClass("FriendOf");
}
And this is the final version of our class
This is the full source code:
import com.orientechnologies.orient.core.db.ODatabaseSession;
import com.orientechnologies.orient.core.db.OrientDB;
import com.orientechnologies.orient.core.db.OrientDBConfig;
public class Main {
public static void main(String[] args) {
OrientDB orient = new OrientDB("remote:localhost", OrientDBConfig.defaultConfig());
ODatabaseSession db = orient.open("test", "admin", "admin");
if (db.getClass("Person") == null) {
db.createVertexClass("Person");
}
if (db.getClass("FriendOf") == null) {
db.createEdgeClass("FriendOf");
}
db.close();
orient.close();
}
}
Now just run it and then open OrientDB Studio again:
- go back to the browser (OrientDB Studio)
- click on the SCHEMA tab
you will see the two newly created classes: Person
and FriendOf
Just to make it a bit more realistic, let's also create some basic schema for our Person
class.
We will just add a name
to the Person and we will create an index on it.
OClass person = db.getClass("Person");
if (person == null) {
person = db.createVertexClass("Person");
}
if (person.getProperty("name") == null) {
person.createProperty("name", OType.STRING);
// index name index type property name
person.createIndex("Person_name_index", OClass.INDEX_TYPE.NOTUNIQUE, "name");
}
And this is the final result:
import com.orientechnologies.orient.core.db.ODatabaseSession;
import com.orientechnologies.orient.core.db.OrientDB;
import com.orientechnologies.orient.core.db.OrientDBConfig;
import com.orientechnologies.orient.core.metadata.schema.OClass;
import com.orientechnologies.orient.core.metadata.schema.OType;
public class Main {
public static void main(String[] args) {
OrientDB orient = new OrientDB("remote:localhost", OrientDBConfig.defaultConfig());
ODatabaseSession db = orient.open("test", "admin", "admin");
OClass person = db.getClass("Person");
if (person == null) {
person = db.createVertexClass("Person");
}
if (person.getProperty("name") == null) {
person.createProperty("name", OType.STRING);
person.createIndex("Person_name_index", OClass.INDEX_TYPE.NOTUNIQUE, "name");
}
if (db.getClass("FriendOf") == null) {
db.createEdgeClass("FriendOf");
}
db.close();
orient.close();
}
}
Now you are ready for the Next Step - Create your First Graph and Run Queries >>>
OrientDB for Java Developers in Five Minutes
In the previous step you established a DB connection from Java, then you created the DB schema (two classes and an index)
Now it's time to insert and query some data
Step 5/5 - Create and Query a Graph
First of all, let's create three vertices: Alice, Bob and Jim
Creating vertices
We are good Java developers, aren't we? Let's encapsulate a single vertex creation in a method:
private static OVertex createPerson(ODatabaseSession db, String name, String surname) {
OVertex result = db.newVertex("Person");
result.setProperty("name", name);
result.setProperty("surname", surname);
result.save();
return result;
}
Wow, we never mentioned that people have a surname
!!! In the previous section we just defined the schema for name
property...
OrientDB can work schemaful (with all the property names and types defined), schemaless (schema-free, no schema defined) or schema-mixed like in this case, where we define a part of the schema (ie. the name
) but we leave the developer the ability to add new properties at run time, without having to deal with further schema definitions.
Now let's create the three vertices:
private static void createPeople(ODatabaseSession db){
OVertex alice = createPerson(db, "Alice", "Foo");
OVertex bob = createPerson(db, "Bob", "Bar");
OVertex jim = createPerson(db, "Jim", "Baz");
}
Creating edges
Suppose that Alice is a friend of Bob and that Bob is a friend of Jim:
Alice --FriendOf--> Bob --FriendOf--> Jim
Let's create the edges in the database:
OEdge edge1 = alice.addEdge(bob, "FriendOf");
edge1.save();
OEdge edge2 = bob.addEdge(jim, "FriendOf");
edge2.save();
Please consider that edges are plain documents, so you can get/set properties on them exactly like for vertices.
Executing queries
Last step of this journey: let's write and execute a simple query that finds friends of friends (FoaF) of a person. We will use a SELECT for this.
private static void executeAQuery(ODatabaseSession db) {
String query = "SELECT expand(out('FriendOf').out('FriendOf')) from Person where name = ?";
OResultSet rs = db.query(query, "Alice");
while (rs.hasNext()) {
OResult item = rs.next();
System.out.println("friend: " + item.getProperty("name"));
}
rs.close(); //REMEMBER TO ALWAYS CLOSE THE RESULT SET!!!
}
or, if you prefer Java Streams API:
private static void executeAQuery(ODatabaseSession db) {
String query = "SELECT expand(out('FriendOf').out('FriendOf')) from Person where name = ?";
OResultSet rs = db.query(query, "Alice");
rs.stream().forEach(x -> System.out.println("friend: " + x.getProperty("name")));
rs.close();
}
Let's try a more complex query, let's find all the people that are friends of both Alice and Jim. We will use a MATCH for this.
private static void executeAnotherQuery(ODatabaseSession db) {
String query =
" MATCH " +
" {class:Person, as:a, where: (name = :name1)}, " +
" {class:Person, as:b, where: (name = :name2)}, " +
" {as:a} -FriendOf-> {as:x} -FriendOf-> {as:b} " +
" RETURN x.name as friend ";
Map<String, Object> params = new HashMap<String, Object>();
params.put("name1", "Alice");
params.put("name2", "Jim");
OResultSet rs = db.query(query, params);
while (rs.hasNext()) {
OResult item = rs.next();
System.out.println("friend: " + item.getProperty("name"));
}
rs.close();
}
Good job!!! This is your first OrientDB Java program!
Here is the full source code of the main class:
import com.orientechnologies.orient.core.db.ODatabaseSession;
import com.orientechnologies.orient.core.db.OrientDB;
import com.orientechnologies.orient.core.db.OrientDBConfig;
import com.orientechnologies.orient.core.metadata.schema.OClass;
import com.orientechnologies.orient.core.metadata.schema.OType;
import com.orientechnologies.orient.core.record.OEdge;
import com.orientechnologies.orient.core.record.OVertex;
import com.orientechnologies.orient.core.sql.executor.OResult;
import com.orientechnologies.orient.core.sql.executor.OResultSet;
import java.util.HashMap;
import java.util.Map;
public class Main {
public static void main(String[] args) {
OrientDB orient = new OrientDB("remote:localhost", OrientDBConfig.defaultConfig());
ODatabaseSession db = orient.open("test", "admin", "admin");
createSchema(db);
createPeople(db);
executeAQuery(db);
executeAnotherQuery(db);
db.close();
orient.close();
}
private static void createSchema(ODatabaseSession db) {
OClass person = db.getClass("Person");
if (person == null) {
person = db.createVertexClass("Person");
}
if (person.getProperty("name") == null) {
person.createProperty("name", OType.STRING);
person.createIndex("Person_name_index", OClass.INDEX_TYPE.NOTUNIQUE, "name");
}
if (db.getClass("FriendOf") == null) {
db.createEdgeClass("FriendOf");
}
}
private static void createPeople(ODatabaseSession db) {
OVertex alice = createPerson(db, "Alice", "Foo");
OVertex bob = createPerson(db, "Bob", "Bar");
OVertex jim = createPerson(db, "Jim", "Baz");
OEdge edge1 = alice.addEdge(bob, "FriendOf");
edge1.save();
OEdge edge2 = bob.addEdge(jim, "FriendOf");
edge2.save();
}
private static OVertex createPerson(ODatabaseSession db, String name, String surname) {
OVertex result = db.newVertex("Person");
result.setProperty("name", name);
result.setProperty("surname", surname);
result.save();
return result;
}
private static void executeAQuery(ODatabaseSession db) {
String query = "SELECT expand(out('FriendOf').out('FriendOf')) from Person where name = ?";
OResultSet rs = db.query(query, "Alice");
while (rs.hasNext()) {
OResult item = rs.next();
System.out.println("friend: " + item.getProperty("name"));
}
rs.close(); //REMEMBER TO ALWAYS CLOSE THE RESULT SET!!!
}
private static void executeAnotherQuery(ODatabaseSession db) {
String query =
" MATCH " +
" {class:Person, as:a, where: (name = :name1)}, " +
" {class:Person, as:b, where: (name = :name2)}, " +
" {as:a} -FriendOf-> {as:x} -FriendOf-> {as:b} " +
" RETURN x.name as friend ";
Map<String, Object> params = new HashMap<String, Object>();
params.put("name1", "Alice");
params.put("name2", "Jim");
OResultSet rs = db.query(query, params);
while (rs.hasNext()) {
OResult item = rs.next();
System.out.println("friend: " + item.getProperty("friend"));
}
rs.close();
}
}
Next steps:
You may be interested in:
Getting Started
Over the past few years, there has been an explosion of many NoSQL database solutions and products. The meaning of the word "NoSQL" is not a campaign against the SQL language. In fact, OrientDB allows for SQL syntax! NoSQL is probably best described by the following:
NoSQL, meaning "not only SQL", is a movement encouraging developers and business people to open their minds and consider new possibilities beyond the classic relational approach to data persistence.
Alternatives to relational database management systems have existed for many years, but they have been relegated primarily to niche use cases such as telecommunications, medicine, CAD and others. Interest in NoSQL alternatives like OrientDB is increasing dramatically. Not surprisingly, many of the largest web companies like Google, Amazon, Facebook, Foursquare and Twitter are using NoSQL based solutions in their production environments.
What motivates companies to leave the comfort of a well established relational database world? It is basically the great need to better solve today's data problems. Specifically, there are a few key areas:
- Performance
- Scalability (often huge)
- Smaller footprint
- Developer productivity and friendliness
- Schema flexibility
Most of these areas also happen to be the requirements of modern web applications. A few years ago, developers designed systems that could handle hundreds of concurrent users. Today it is not uncommon to have a potential target of thousands or millions of users connected and served at the same time.
Changing technology requirements have been taken into account on the application front by creating frameworks, introducing standards and leveraging best practices. However, in the database world, the situation has remained more or less the same for over 30 years. From the 1970s until recently, relational DBMSs have played the dominant role. Programming languages and methodologies have evolved, but the concept of data persistence and the DBMS have remained unchanged for the most part: it is all still tables, records and joins.
NoSQL Models
NoSQL-based solutions in general provide a powerful, scalable, and flexible way to solve data needs and use cases, which have previously been managed by relational databases. To summarize the NoSQL options, we'll examine the most common models or categories:
-
Key / Value databases: where the data model is reduced to a simple hash table, which consists of key / value pairs. It is often easily distributed across multiple servers. The most recognized products of this group include Redis, Dynamo, and Riak.
-
Column-oriented databases: where the data is stored in sections of columns offering more flexibility and easy aggregation. Facebook's Cassandra, Google's BigTable, and Amazon's SimpleDB are some examples of column-oriented databases.
-
Document databases: where the data model consists of document collections, in which each individual document can have multiple fields without necessarily having a defined schema. The best known products of this group are MongoDB and CouchDB.
-
Graph databases: where the domain model consists of vertices interconnected by edges creating rich graph structures. The best known products of this group are OrientDB, Neo4j and Titan.
OrientDB is a document-graph database, meaning it has full native graph capabilities coupled with features normally only found in document databases.
Each of these categories or models has its own peculiarities, strengths and limitations. There is no single category or model, which is better than the others. However, certain types of databases are better at solving specific problems. This leads to the motto of NoSQL: choose the best tool for your specific use case.
The goal of Orient Technologies in building OrientDB was to create a robust, highly scalable database that can perform optimally in the widest possible set of use cases. Our product is designed to be a fantastic "go to" solution for practically all of your data persistence needs. In the following parts of this tutorial, we will look closely at OrientDB, one of the best open-source, multi-model, next generation NoSQL products on the market today.
Installation
For a complete installation guide, see OrientDB Server Administration - Installation
Running the OrientDB Server
When you finish installing OrientDB, whether you build it from source or download the binary package, you are ready to launch the database server. You can either start it through the system daemon or through the provided server script. This article only covers the latter.
Note: If you would like to run OrientDB as a service on your system, there are some additional steps that you need to take. This provides alternate methods for starting the server and allows you to launch it as a daemon when your system boots.
Note that you need to start the OrientDB server once manually via server.sh|bat in %ORIENTDB_HOME%\bin once, before starting the service.
For more information on this process see:
Starting the Database Server
While you can run the database server as system daemon, you also have the option of starting it directly. In the OrientDB installation directory, (that is $ORIENTDB_HOME
), under bin
, there is a file named server.sh
on Unix-based systems and server.bat
on Windows. Executing this file starts the server.
To launch the OrientDB database server, run the following commands:
$cd $ORIENTDB_HOME/bin
$./server.sh
. .` ` , `:. `,` ,:` .,. :,, .,, ,,, . .,.::::: ```` ,` .::,,,,::.,,,,,,`;; .: `,. ::,,,,,,,:.,,.` ` .: ,,:,:,,,,,,,,::. ` ` `` .: ,,:.,,,,,,,,,: `::, ,, ::,::` : :,::` :::: ,:,,,,,,,,,,::,: ,, :. : :: : .: :,,,,,,,,,,:,:: ,, : : : : .: ` :,,,,,,,,,,:,::, ,, .:::::::: : : .: `,...,,:,,,,,,,,,: .:,. ,, ,, : : .: .,,,,::,,,,,,,: `: , ,, : ` : : .: ...,::,,,,::.. `: .,, :, : : : .: ,::::,,,. `: ,, ::::: : : .: ,,:` `,,. ,,, .,` ,,. `, S E R V E R `` `. `` ` 2012-12-28 01:25:46:319 INFO Loading configuration from: config/orientdb-server-config.xml... [OServerConfigurationLoaderXml] 2012-12-28 01:25:46:625 INFO OrientDB Server v1.6 is starting up... [OServer] 2012-12-28 01:25:47:142 INFO -> Loaded memory database 'temp' [OServer] 2012-12-28 01:25:47:289 INFO Listening binary connections on 0.0.0.0:2424 [OServerNetworkListener] 2012-12-28 01:25:47:290 INFO Listening http connections on 0.0.0.0:2480 [OServerNetworkListener] 2012-12-28 01:25:47:317 INFO OrientDB Server v1.6 is active. [OServer]
The database server is now running. It is accessible on your system through ports 2424
and 2480
.
At the first startup the server will ask for the root user password. The password is stored in the config file.
Stop the Server
On the console where the server is running a simple CTRL+c will shutdown the server.
The shutdown.sh (shutdown.bat) script could be used to stop the server:
$cd $ORIENTDB_HOME/bin
$./shutdown.sh -p ROOT_PASSWORD
On *nix systems a simple call to shutdown.sh will stop the server running on localhost:
$cd $ORIENTDB_HOME/bin
$./shutdown.sh
It is possible to stop servers running on remote hosts or even on different ports on localhost:
$cd $ORIENTDB_HOME/bin
$./shutdown.sh -h odb1.mydomain.com -P 2424-2430 -u root -p ROOT_PASSWORD
List of params
- -h | --host HOSTNAME or IP ADDRESS : the host or ip where OrientDB is running, default to localhost
- -P | --ports PORT or PORT RANGE : single port value or range of ports; default to 2424-2430
- -u | --user ROOT USERNAME : root's username; default to root
- -p | --password ROOT PASSWORD : root's user password; mandatory
NOTE: On Windows systems password is always mandatory because the script isn't able to discover the pid of the OrientDB's process.
Server Log Messages
Following the masthead, the database server begins to print log messages to standard output. This provides you with a guide to what OrientDB does as it starts up on your system.
-
The database server loads its configuration file from the file
$ORIENTDB_HOME/config/orientdb-server-config.xml
.For more information on this step, see OrientDB Server.
-
The database server loads the
temp
database into memory. You can use this database for storing temporary data. -
The database server begins listening for binary connections on port
2424
for all configured networks, (0.0.0.0
). -
The database server begins listening for HTTP connections on port
2480
for all configured networks, (0.0.0.0
).
Accessing the Database Server
By default, OrientDB listens on two different ports for external connections.
-
Binary: OrientDB listens on port
2424
for binary connections from the console and for clients and drivers that support the Network Binary Protocol. -
HTTP: OrientDB listens on port
2480
for HTTP connections from OrientDB Studio Web Tool and clients and drivers that support the HTTP/REST protocol, or similar tools, such as cURL.
If you would like the database server to listen at different ports or IP address, you can define these values in the configuration file config/orientdb-server-config.xml
.
Running the OrientDB Console
Once the server is running there are various methods you can use to connect to your database server to an individual databases. Two such methods are the Network Binary and HTTP/REST protocols. In addition to these OrientDB provides a command-line interface for connecting to and working with the database server.
Starting the OrientDB Console
In the OrientDB installation directory (that is, $ORIENTDB_HOME
, where you installed the database) under bin
, there is a file called console.sh
for Unix-based systems or console.bat
for Windows users.
To launch the OrientDB console, run the following command after you start the database server:
$cd $ORIENTDB_HOME/bin
$./console.sh
OrientDB console v.X.X.X (build 0) www.orientdb.com Type 'HELP' to display all the commands supported. Installing extensions for GREMLIN language v.X.X.X orientdb>
The OrientDB console is now running. From this prompt you can connect to and manage any remote or local databases available to you.
Using the HELP
Command
In the event that you are unfamiliar with OrientDB and the available commands, or if you need help at any time, you can use the HELP
command, or type ?
into the console prompt.
orientdb> HELP
AVAILABLE COMMANDS:
* alter class <command-text> Alter a class in the database schema
* alter cluster <command-text> Alter class in the database schema
... ...
* help Print this help
* exit Close the console
For each console command available to you, HELP
documents its basic use and what it does. If you know the particular command and need details on its use, you can provide arguments to HELP
for further clarification.
orientdb> HELP SELECT
COMMAND: SELECT
- Execute a query against the database and display the results.
SYNTAX: select <query-text>
WHERE:
- <query-text>: The query to execute
Connecting to Server Instances
There are some console commands, such as LIST DATABASES
or CREATE DATABASE
, which you can only run while connected to a server instance. For other commands, however, you must also connect to a database, before they run without error.
Before you can connect to a fresh server instance and fully control it, you need to know the root password for the database.
With the required credentials, you can connect to the database server instance on your system, or establish a remote connection to one running on a different machine.
orientdb> CONNECT remote:localhost root my_root_password
Connecting to remote Server instance [remote:localhost] with user 'root'...OK
Once you have established a connection to the database server, you can begin to execute commands on that server, such as LIST DATABASES
and CREATE DATABASE
.
orientdb> LIST DATABASES
Found 1 databases:
* GratefulDeadConcerts (plocal)
To connect to this database or to a different one, use the CONNECT
command from the console and specify the server URL, username, and password. By default, each database has an admin
user with a password of admin
.
Warning: Always change the default password on production databases.
The above LIST DATABASES
command shows a GratefulDeadConcerts
installed on the local server. To connect to this database, run the following command:
orientdb> CONNECT remote:localhost/GratefulDeadConcerts admin admin
Connecting to database [remote:localhost/GratefulDeadConcerts] with user 'admin'...OK
The CONNECT
command takes a specific syntax for its URL. That is, remote:localhost/GratefulDeadConcerts
in the example. It has three parts:
-
Protocol: The first part of the database address is the protocol the console should use in the connection. In the example, this is
remote
, indicating that it should use the TCP/IP protocol. -
Address: The second part of the database address is hostname or IP address of the database server that you want the console to connect to. In the example, this is
localhost
, since the connection is made to a server instance running on the local file system. -
Database: The third part of the address is the name of the database that you want to use. In the case of the example, this is
GratefulDeadConcerts
.
For more detailed information about the commands, see Console Commands.
Note: The OrientDB distribution comes with the bundled database
GratefulDeadConcerts
which represents the Graph of the Grateful Dead's concerts. This database can be used by anyone to start exploring the features and characteristics of OrientDB.
Run the Studio
If you're more comfortable interacting with database systems through a graphical interface then you can accomplish the most common database tasks with OrientDB Studio, the web interface.
Connecting to Studio
By default, there are no additional steps that you need to take to start OrientDB Studio. When you launch the Server, whether through the start-up script server.sh
or as a system daemon, the Studio web interface opens automatically with it.
$ firefox http://localhost:2480
From here you can create a new database, connect to or drop an existing database, import a public database and navigate to the Server management interface.
For more information on the OrientDB Studio, see Studio.
Documents, Vertices and Edges
###TODO
Classes
Here we will learn about how classes structure data in OrientDB. A class in OrientDB is similar to a table in a relational database with some key differences. In this section you will learn how to see all of the classes in your database and how to create classes of your own. You'll also learn how to provide schema on top of classes by defining constraints for a class's properties. Finally you'll learn how to access the records stored within your classes.
The Class is a concept drawn from the Object-oriented programming paradigm. In OrientDB a class is a data model that allows you to define certain rules for records that belong together. For example, a class 'Person' can store information about people. You can structure your class such that a record in the class must have certain properties (i.e. Name, Birthdate, Favorite Number, etc...).
In the traditional document database model classes are comparable to collections, while in the Relational database model (R-DBMS) they are comparable to tables. Classes are not tables though. Classes provide efficient means for storage of schema-less data. We'll see more about schema-less, schema-full, and schema-mixed data models later (See 'Adding Properties to a Class' below).
Like many database management systems, OrientDB uses the Record as an element of storage. There are many types of records, but with the Document Database API records always use the Document type. A document is formed by a set of key/value pairs. A document can belong to a class.
In the example above, there are two documents. One document contains information for Sarah and another for Abdul. The keys 15:0 and 15:1 refer to each document respectively.
To list all the configured classes on your system, use the LIST CLASSES
command in the console:
orientdb> LIST CLASSES
orientdb {db=playingwithClasses}> LIST CLASSES
CLASSES
+----+-----------+-------------+-----------------+-----+
|# |NAME |SUPER-CLASSES|CLUSTERS |COUNT|
+----+-----------+-------------+-----------------+-----+
|0 |_studio | |_studio(13) | 1|
|1 |Blue |[Color] |blue(19) | 0|
|2 |Color |[V] |- | 0|
|3 |E | |e(11),e_1(12) | 0|
|4 |OFunction | |ofunction(6) | 0|
|5 |OIdentity | |- | 0|
|6 |ORestricted| |- | 0|
|7 |ORole |[OIdentity] |orole(4) | 3|
|8 |OSchedule | |oschedule(8) | 0|
|9 |OSequence | |osequence(7) | 0|
|10 |OTriggered | |- | 0|
|11 |OUser |[OIdentity] |ouser(5) | 3|
|12 |Person |[V] |person(15) | 0|
|13 |Red |[Color] |red(17),red_1(18)| 0|
|14 |V | |v(9),v_1(10) | 0|
+----+-----------+-------------+-----------------+-----+
| |TOTAL | | | 7|
+----+-----------+-------------+-----------------+-----+
If you are using studio, then you can see the same information by clicking on the 'schema' tab.
Here we can see that there are 14 classes in the database. Class 12 refers to person. There is also a class Color which is the super-class of Red and Blue. Color and Person both have super-classes called V. The class V is important for using OrientDB’s graph model. We'll see more about Superclasses and V later in the tutorial. Let's move on now to working with classes.
Working with Classes
In order to start using classes with your own applications, you need to understand how to create and configure a class for use. The class in OrientDB is similar to the table in relational databases, but unlike tables, classes can be schema-less, schema-full or mixed. A class can inherit properties from other classes thereby creating trees of classes (though the super-class relationship).
Each class has its own cluster or clusters, (created by default, if none are defined). For now we should know that a cluster is a place where a group of records are stored. We'll soon see how clustering
improves performance of querying the database.
For more information on classes in OrientDB, see Class.
To create a new class, use the CREATE CLASS
command:
orientdb> CREATE CLASS Student
Class created successfully. Total classes in database now: 15
This creates a class called Student
. Given that no cluster was defined in the CREATE CLASS
command, OrientDB creates a default cluster called student
, to contain records assigned to this class. For the moment, the class has no records or properties tied to it. It is now displayed in the CLASSES
listing and in the schema manager of Studio.
Adding Properties to a Class
As mentioned above, OrientDB allows you to work in a schema-less mode. That is, it allows you to create classes without defining their properties. However, properties are mandatory if you would like to define indexes or constraints for a class. Let's follow OrientDB's comparison to relational databases again... If classes in OrientDB are similar to tables, then properties are the columns on those tables.
To create new properties on Student
, use the CREATE PROPERTY
command in the console or in the browse window of studio:
orientdb>CREATE PROPERTY Student.name STRING
Property created successfully with id=1 orientdb>CREATE PROPERTY Student.surname STRING
Property created successfully with id=2 orientdb>CREATE PROPERTY Student.birthDate DATE
Property created successfully with id=3
These commands create three new properties on the Student
class. The properties provide you with areas to define an individual student's name, surname, and date of birth.
Displaying Class Information
Occasionally you may need to reference a particular class to see what clusters it belongs to, or any properties configured for the class's use. Use the INFO CLASS
command to display information about the current configuration and properties of a class.
To display information on the class Student
, use the INFO CLASS
command:
orientdb> INFO CLASS Student
Class................: Student
Default cluster......: student (id=96)
Supported cluster ids: [96]
Properties:
-----------+--------+--------------+-----------+----------+----------+-----+-----+
NAME | TYPE | LINKED TYPE/ | MANDATORY | READONLY | NOT NULL | MIN | MAX |
| | CLASS | | | | | |
-----------+--------+--------------+-----------+----------+----------+-----+-----+
birthDate | DATE | null | false | false | false | | |
name | STRING | null | false | false | false | | |
surname | STRING | null | false | false | false | | |
-----------+--------+--------------+-----------+----------+----------+-----+-----+
Adding Constraints to Properties
Constraints create limits on the data values assigned to properties. For instance, the type, the minimum or maximum size of, whether or not a value is mandatory or if null values are permitted to the property.
Constraints create limits on the data values assigned to properties. For instance, if 'MANDATORY' is set to true for name in student, then every record in the student class must have a name. If we set 'MIN' to three, then every name must also be at least three characters long.
The only two properties required when using the 'create a property' command for a class are 'NAME' and 'TYPE'.
To add a constraint to an existing property, use the ALTER PROPERTY
command:
orientdb> ALTER PROPERTY Student.name MIN 3
Property updated successfully
This command adds a constraint to Student
on the name
property. After running this command, Student will allow any record to be stored unless the record has a property called 'Name'. If the records has such a property then 'Student' will reject the record if the value in 'Name' is less then three characters.
By setting property, 'MANDATORY', to true for Student's Name we can also guarantee that every record added to student has a name.
orientdb> ALTER PROPERTY Student.name MANDATORY true
There are many ways to use constraints on properties. They can allow you to build a data-model that tells a story about your own use case. Constraints can also help ensure that you're database communicates with other components of a larger application by only allowing storage of values that another application is able to recognize.
Viewing Records in a Class
Classes contain and define records in OrientDB. You can view all records that belong to a class using the BROWSE CLASS
command. You can also see data belonging to a particular record with the DISPLAY RECORD
command.
Note: you cannot display a record unless you have recently received a query result with records to browse (select statement, 'browse class x', etc...).
Earlier we created a Student
class and defined some schema for records belonging to that class, but we didn't create any records or add any data. Thus, running 'BROWSE CLASS' on the Student
class returns no results. Luckily OrientDB has a few preconfigured classes and records that we can query.
Let's take the class OUser for example.
orientdb> INFO CLASS OUser
CLASS 'OUser'
Super classes........: [OIdentity]
Default cluster......: ouser (id=5)
Supported cluster ids: [5]
Cluster selection....: round-robin
Oversize.............: 0.0
PROPERTIES
----------+---------+--------------+-----------+----------+----------+-----+-----+
NAME | TYPE | LINKED TYPE/ | MANDATORY | READONLY | NOT NULL | MIN | MAX |
| | CLASS | | | | | |
----------+---------+--------------+-----------+----------+----------+-----+-----+
password | STRING | null | true | false | true | | |
roles | LINKSET | ORole | false | false | false | | |
name | STRING | null | true | false | true | | |
status | STRING | null | true | false | true | | |
----------+---------+--------------+-----------+----------+----------+-----+-----+
INDEXES (1 altogether)
-------------------------------+----------------+
NAME | PROPERTIES |
-------------------------------+----------------+
OUser.name | name |
-------------------------------+----------------+
The OUser
class defines the users on your database.
To see records assigned to the OUser
class, run the BROWSE CLASS
command:
orientdb> BROWSE CLASS OUser
---+------+-------+--------+-----------------------------------+--------+-------+
# | @RID | @Class| name | password | status | roles |
---+------+-------+--------+-----------------------------------+--------+-------+
0 | #5:0 | OUser | admin | {SHA-256}8C6976E5B5410415BDE90... | ACTIVE | [1] |
1 | #5:1 | OUser | reader | {SHA-256}3D0941964AA3EBDCB00EF... | ACTIVE | [1] |
2 | #5:2 | OUser | writer | {SHA-256}B93006774CBDD4B299389... | ACTIVE | [1] |
---+------+-------+--------+-----------------------------------+--------+-------+
In the example, you are listing all of the users of the database. While this is fine for your initial setup and as an example, it is not particularly secure. To further improve security in production environments, see Security. |
When you run BROWSE CLASS
, the first column in the output provides the identifier number, which you can use to display detailed information on that particular record.
To show the first record browsed from the OUser
class, run the DISPLAY RECORD
command:
orientdb> DISPLAY RECORD 0
DOCUMENT @class:OUser @rid:#5:0 @version:1
----------+--------------------------------------------+
Name | Value |
----------+--------------------------------------------+
name | admin |
password | {SHA-256}8C6976E5B5410415BDE908BD4DEE15... |
status | ACTIVE |
roles | [#4:0=#4:0] |
----------+--------------------------------------------+
Bear in mind that this command references the last call of BROWSE CLASS
. You can continue to display other records, but you cannot display records from another class until you browse that particular class.
Class Review
Here are some key things to remember about classes:
-
A class in OrientDB is similar to a table in a relational database with some key differences. Among those differences we see tables are schema-full, and classes can be schema-full, schema-less, or mixed.
-
You can see all of the classes in your database by running 'LIST CLASSES' in console or by visiting the 'Schema Manager' in Studio.
-
You can create a class by running the 'create class
' command in console, or by running the same command in the 'Browse' window of studio. -
You can use the commands, 'Create property
[constraints]' and 'Create property [constraints]' to give schema to a class. -
To see properties and constraints associated with a class you can run 'info class
'. -
To see information about a the records within a class run 'Browse class
'. -
To see information about a specific record of a class use the command 'Display record
'. Note: You must have recently queried a class for it's records before using this command. ' ' references the number in the left-most column of the previous query's result.
Congratulations! You are now familiar with classes in OrientDB. If you're ready to explore clusters then let's move on to the clustering
section of this tutorial.
Demo Database
Starting with OrientDB v.3.0 a new demo database (demodb
) is included.
This Section introduces the demodb
database and its Data Model, and includes some queries that is possible to execute on it.
Some screenshots on "portions" of the demodb
graph are included as well.
Introduction
The demo database can help you understand better OrientDB features and capabilities and replaces the old demo database GratefulDeadConcerts
included in version 2.2 and previous ones.
Note: Random-generated data is used in the demodb
, including data used for Emails, Names, Surnames, Phone Numbers and Reviews.
Version
demodb
has a version that, in general, is not linked to the Server version you are running. You can check the version of the demodb
included in your distribution by executing the following SQL query:
SELECT `Version` FROM `DBInfo`;
Current version is 0.76.
Location
The demo database is located in the databases
directory under your $ORIENTDB_HOME
(e.g. D:\orientdb\orientdb-community-3.0.0\databases\demodb
.
Connecting to the Demo Database
It is possible to connect to demodb
using the three standard OrientDB Users:
reader
/reader
writer
/writer
admin
/admin
Using the Demo Database with OrientDB 2.2.x
The demo database can be easily loaded into OrientDB 2.2.x using the Studio's "Import a public database" feature.
Alternatively, it is possible to import it via an SQL script that includes the set of instructions needed to recreate the data model as well as all the records (vertices, edges, and documents).
Data Model
demodb
is the database of an hypothetical Travel Agency that runs a public social platform as well.
Users (that are stored in the database in the class Profiles) can freely register to the social platform and start making friends (friendship is expressed via the HasFriend edge).
Some of the users can become customers. When this happens the application in use at the Social Travel Agency creates a vertex in the Customers class and links it to the associated profile via an HasProfile edge.
When customers are created, they are automatically linked to a country as well, via an IsFromCountry edge. Countries are stored in the Countries vertex class.
Orders made by customers are stored in the vertex class Orders. Each customer can make one or more orders, and the HasCustomer edge is used to connect orders to customers.
When customers start visiting Attractions (like Castles, Monuments, Theatres or Archaeological Sites) or using Services (like Hotels or Restaurants) edges are created to link that specific customer with that specific attraction or service (HasVisited, HasStayed, and HasEaten edges are used).
The Social Travel Agency also stores some reviews in the vertex class Reviews. Reviews are linked to customers via the MadeReview edge, and to an attraction or service via the HasReview edge.
Data model of demodb
v.0.76 is reported in the image below:
Inheritance
Inheritance in the Vertex and Edge classes of demodb
v.0.76 is reported in the image below:
Queries
This Section includes several query examples that you can execute from the Studio's Browse Tab, or from its Graph Editor. You may also execute these queries directly from the Console, or your application through an API or Driver.
{% include "../include-demodb-version-warning.md" %}
The following table can help you navigate through all examples:
Category | Question | Link |
---|---|---|
Profiles | Find the 'year of birth' of the Profiles, and how many Profiles were born in the same year | Link |
Find the top 3 Profiles that have the highest number of Friends | Link | |
Friendship | Find Santo's Friends | Link |
Find Santo's Friends who are also Customers | Link | |
Find Santo's Friends who are also Customers, and the Countries they are from | Link | |
Find Santo's Friends who are also Customers, and the Orders they have placed | Link | |
Among Santo's Friends, find the top 3 Customers that placed the highest number of Orders | Link | |
Among Santo's Friends, find the top 3 Customers that visited the highest number of Places | Link | |
Find all the Friends of Customer identified with OrderedId 1 that are not Customers (so that a product can be proposed) | Link | |
Customers | Find everything that is connected (1st degree) to Customer with OrderedId 1 | Link |
Find all Locations connected to Customer with OrderedId 1 | Link | |
Find all Locations connected to Customer with OrderedId 1, and their Reviews (if any) | Link | |
Find the other Customers that visited the Locations visited by Customer with OrderedId 1 | Link | |
Find all the places where Customer with OrderedId 2 has stayed | Link | |
Find all places where Customer with Id 1 has eaten | Link | |
Find the 3 Customers who made more reviews | Link | |
Find all Orders placed by Customer with OrderedId 2 | Link | |
Calculate the total revenues from Orders associated with Customer with OrderedId 2 | Link | |
Find the 3 Customers who placed most Orders | Link | |
Find the top 3 Countries from where Customers are from | Link | |
Countries | Find the top 3 Countries from where Customers are from | Link |
Find Santo's Friends who are also Customers, and the Countries they are from | Link | |
Orders | Calculate the total revenues from Orders, and the min, max and average Order amount | Link |
Find the year of the Orders, and how many Orders have been placed in the same year | Link | |
Find the 3 Customers who placed most Orders | Link | |
Find the top 3 Customers in terms of spending | Link | |
Find all Orders placed by Customer with Id 2 | Link | |
Calculate the total revenues from Orders associated with Customer with Id 2 | Link | |
Attractions | Find all Attractions connected with Customer with OrderedId 1 | Link |
Services | Find all Services connected with Customer with OrderedId 1 | Link |
Find the 3 Hotels that have been booked most times | Link | |
Find the 3 Hotels that have most reviews | Link | |
Find the top 3 nationality of the tourists that have eaten at Restaurant with Id 26 | Link | |
Locations | Find all Locations visited by Santo | Link |
Find all Locations connected to Customer with OrderedId 1 | Link | |
Find all Locations connected to Customer with OrderedId 1, and their Reviews (if any) | Link | |
Find all Locations visited by Customer with OrderedId 2 | Link | |
Find all Locations visited by Santo's friends | Link | |
Reviews | Find number of Reviews per star | Link |
Find all reviewed Services | Link | |
Find all reviewed Services and the Customer who made the review | Link | |
Find the numbers of reviews per Service | Link | |
Find the 3 Hotels that have most reviews | Link | |
Find the 3 Customers who made more reviews | Link | |
Recommendations | Recommend some friends to Profile 'Isabella Gomez' (friends of friends) | Link |
Recommend some Hotels to Customer with OrderedId 1 | Link | |
Business Opportunities | Find all the Friends of Customer identified with OrderedId 1 that are not Customers (so that a product can be proposed) | Link |
Find all the Customer Friends that are not Customers (so that a product can be proposed) | Link | |
Polymorphism | Find all Locations (Services + Attractions) connected with Customer with OrderedId 1 | Link |
Find the 3 Services (Hotels + Restaurants) that have most reviews | Link | |
Shortest Paths | Find the shortest path between the Profile 'Santo' and the Country 'United States' | Link |
Find the shortest path between the Profile 'Santo' and the Restaurant 'Malga Granezza' | Link | |
Traverses | Traverse everything from Profile 'Santo' up to depth three | Link |
Traverse everything from Country 'Italy' up to depth three | Link |
Profiles
Example 1
Find the 'year of birth' of the Profiles, and how many Profiles were born in the same year:
SELECT
count(*) as NumberOfProfiles,
Birthday.format('yyyy') AS YearOfBirth
FROM Profiles
GROUP BY YearOfBirth
ORDER BY NumberOfProfiles DESC
In the Browse Tab of Studio, using the query above, this is the obtained list of records (only few records are shown in the image below):
Note: in v. 2.x the corresponding query is:
SELECT
count(*) as NumberOfProfiles,
YearOfBirth
FROM (
SELECT
Birthday.format('yyyy') AS YearOfBirth
FROM Profiles)
GROUP BY YearOfBirth
ORDER BY NumberOfProfiles DESC
Example 2
Find the top 3 Profiles that have the highest number of Friends:
SELECT
@rid as Profile_RID,
Name,
Surname,
both('HasFriend').size() AS FriendsNumber
FROM `Profiles`
ORDER BY FriendsNumber DESC
LIMIT 3
In the Browse Tab of Studio, using the query above, this is the obtained list of records:
Friendship
Example 1
Find Santo's Friends:
MATCH {Class: Profiles, as: profile, where: (Name='Santo' AND Surname='OrientDB')}-HasFriend-{Class: Profiles, as: friend}
RETURN $pathelements
In the Graph Editor included in Studio, using 'RETURN $pathelements' as RETURN
clause, this is the obtained graph:
In the Browse Tab of Studio, using 'RETURN friend.@Rid as Friend_RID, friend.Name as Friend_Name, friend.Surname as Friend_Surname' as RETURN
clause, this is the obtained list of records (only few records are shown in the image below):
If you would like only to count them, you can execute a query like the following:
SELECT COUNT(*)
FROM (
MATCH {Class: Profiles, as: profile, where: (Name='Santo' AND Surname='OrientDB')}-HasFriend-{Class: Profiles, as: friend}
RETURN friend
)
or
SELECT
both('HasFriend').size() AS FriendsNumber
FROM `Profiles`
WHERE Name='Santo' AND Surname='OrientDB'
Example 2
Find Santo's Friends who are also Customers:
MATCH {Class: Profiles, as: profile, where: (Name='Santo' AND Surname='OrientDB')}-HasFriend-{Class: Profiles, as: friend}<-HasProfile-{class: Customers, as: customer}
RETURN $pathelements
In the Graph Editor included in Studio, using 'RETURN $pathelements' as RETURN
clause, this is the obtained graph:
In the Browse Tab of Studio, using 'RETURN friend.@Rid as Friend_RID, friend.Name as Friend_Name, friend.Surname as Friend_Surname, customer.@Rid as Customer_RID, customer.OrderedId as Customer_OrderedId' as RETURN
clause, this is the obtained list of records (only few records are shown in the image below):
Example 3
{% include "./include-file-5.md" %}
Example 4
Find Santo's Friends who are also Customers, and the Orders they have placed:
MATCH {Class: Profiles, as: profile, where: (Name='Santo' AND Surname='OrientDB')}-HasFriend-{Class: Profiles, as: friend}<-HasProfile-{class: Customers, as: customer}<-HasCustomer-{Class: Orders, as: order}
RETURN $pathelements
In the Graph Editor included in Studio, using 'RETURN $pathelements' as RETURN
clause, this is the obtained graph:
In the Browse Tab of Studio, using 'RETURN friend.@Rid as Friend_RID, friend.Name as Friend_Name, friend.Surname as Friend_Surname, customer.@Rid as Customer_RID, customer.OrderedId as Customer_OrderedId, order.Id as OrderId' as RETURN
clause, this is the obtained list of records (only few records are shown in the image below):
Example 5
Among Santo's Friends, find the top 3 Customers that placed the highest number of Orders:
SELECT
OrderedId as Customer_OrderedId,
in('HasCustomer').size() as NumberOfOrders,
out('HasProfile').Name as Friend_Name,
out('HasProfile').Surname as Friend_Surname
FROM (
SELECT expand(customer)
FROM (
MATCH {Class: Profiles, as: profile, where: (Name='Santo' AND Surname='OrientDB')}-HasFriend-{Class: Profiles, as: friend}<-HasProfile-{class: Customers, as: customer}
RETURN customer
)
)
ORDER BY NumberOfOrders DESC
LIMIT 3
In the Browse Tab of Studio, using the query above, this is the obtained list of records:
Example 6
Among Santo's Friends, find the top 3 Customers that visited the highest number of Places:
SELECT
OrderedId as Customer_OrderedId,
out('HasVisited').size() as NumberOfVisits,
out('HasProfile').Name as Friend_Name,
out('HasProfile').Surname as Friend_Surname
FROM (
SELECT expand(customer)
FROM (
MATCH {Class: Profiles, as: profile, where: (Name='Santo' AND Surname='OrientDB')}-HasFriend-{Class: Profiles, as: friend}<-HasProfile-{class: Customers, as: customer}
RETURN customer
)
)
ORDER BY NumberOfVisits DESC
LIMIT 3
In the Browse Tab of Studio, using the query above, this is the obtained list of records:
You may find in a similar way the top 3 Customers, among Santo's Friends, that have stayed at the highest number of Hotels, or have eaten at the highest number of Restaurants. Just use out("HasStayed").size()
or out("HasEaten").size()
instead of out("HasVisited").size()
(you may also consider to modify the alias, from NumberOfVisits
to NumberOfHotels
and NumberOfRestaurants
, so that it is more coherent to these cases).
Example 7
{% include "./include-file-1.md" %}
or, without restricting to a specific customer:
Find all the Customer Friends that are not Customers (so that a product can be proposed):
In the Graph Editor included in Studio, using the query below, this is the obtained graph:
SELECT *
FROM (
SELECT expand(customerFriend)
FROM (
MATCH {Class:Customers, as: customer}-HasProfile-{Class:Profiles, as: profile}-HasFriend-{Class:Profiles, as: customerFriend}
RETURN customerFriend
)
)
WHERE in('HasProfile').size()=0
In the Browse Tab of Studio, using the query below, this is the obtained list of records (only few records are shown in the image below):
SELECT
@Rid as Friend_RID,
Name as Friend_Name,
Surname as Friend_Surname
FROM (
SELECT expand(customerFriend)
FROM (
MATCH {Class:Customers, as: customer}-HasProfile-{Class:Profiles, as: profile}-HasFriend-{Class:Profiles, as: customerFriend}
RETURN customerFriend
)
)
WHERE in('HasProfile').size()=0
Customers
Example 1
Find everything that is connected (1st degree) to Customer with OrderedId 1:
MATCH {class: Customers, as: c, where: (OrderedId=1)}--{as: n}
RETURN $pathelements
In the Graph Editor included in Studio, using 'RETURN $pathelements' as RETURN
clause, this is the obtained graph:
Example 2
{% include "./include-file-7.md" %}
Example 3
{% include "./include-file-8.md" %}
Example 4
Find the other Customers that visited the Locations visited by Customer with OrderedId 1:
MATCH {class: Customers, as: customer, where: (OrderedId=1)}--{class: Locations, as: loc}--{class: Customers, as: otherCustomers, where: (OrderedId<>1)}
RETURN $pathelements
In the Graph Editor included in Studio, using 'RETURN $pathelements' as RETURN
clause, this is the obtained graph:
If we want to return also also their Profile names, surnames and emails:
MATCH {class: Customers, as: customer, where: (OrderedId=1)}--{class: Locations, as: loc}--{class: Customers, as: otherCustomers, where: (OrderedId<>1)}-HasProfile->{class: Profiles, as: profile}
RETURN otherCustomers.OrderedId, profile.Name, profile.Surname, profile.Email
ORDER BY `otherCustomers.OrderedId` ASC
In the Browse Tab of Studio, using 'RETURN otherCustomers.OrderedId, profile.Name, profile.Surname, profile.Email' as RETURN
clause, this is the obtained list of records (only few records are shown in the image below):
Example 5
Find all the places where Customer with OrderedId 2 has stayed:
MATCH {as: n}<-HasStayed-{class: Customers, as: c, where: (OrderedId=2)}
RETURN $pathelements
In the Graph Editor included in Studio, using 'RETURN $pathelements' as RETURN
clause, this is the obtained graph:
Example 6
Find all places where Customer with OrderedId 1 has eaten:
MATCH {as: n}<-HasEaten-{class: Customers, as: c, where: (OrderedId=1)}
RETURN $pathelements
In the Graph Editor included in Studio, using 'RETURN $pathelements' as RETURN
clause, this is the obtained graph:
Example 7
{% include "./include-file-3.md" %}
Example 8
{% include "./include-file-9.md" %}
Example 9
{% include "./include-file-10.md" %}
Example 10
{% include "./include-file-4.md" %}
Example 11
{% include "./include-file-6.md" %}
Countries
Example 1
{% include "./include-file-6.md" %}
Example 2
{% include "./include-file-5.md" %}
Orders
Example 1
Calculate the total revenues from Orders, and the min, max and average Order amount:
SELECT
count(*) as OrdersNumber,
sum(Amount) AS TotalRevenuesFromOrders,
min(Amount) as MinAmount,
(sum(Amount)/count(*)) as AverageAmount,
max(Amount) as MaxAmount
FROM Orders
In the Browse Tab of Studio, using the query above, this is the visualized result:
Example 2
Find the year of the Orders, and how many Orders have been placed in the same year:
SELECT
count(*) as OrdersCount,
OrderDate.format('yyyy') AS OrderYear
FROM Orders
GROUP BY OrderYear
ORDER BY OrdersCount DESC
In the Browse Tab of Studio, using the query above, this is the visualized result:
Example 3
{% include "./include-file-4.md" %}
Example 4
Find the top 3 Customers in terms of spending:
SELECT
customer.OrderedId as customerOrderedId,
SUM(order.Amount) as totalAmount
FROM (
MATCH {Class: Customers, as: customer}<-HasCustomer-{class: Orders, as: order}
RETURN customer, order
)
GROUP BY customerOrderedId
ORDER BY totalAmount DESC
LIMIT 3
In the Browse Tab of Studio, using the query above, this is the visualized result:
Example 5
{% include "./include-file-9.md" %}
Example 6
{% include "./include-file-10.md" %}
Attractions
Example 1
Find all Attractions connected with Customer with OrderedId 1:
MATCH {class: Customers, as: customer, where: (OrderedId=1)}--{Class: Attractions, as: attraction}
RETURN $pathelements
In the Graph Editor included in Studio, using 'RETURN $pathelements' as RETURN
clause, this is the obtained graph:
In the Browse Tab of Studio, using 'RETURN attraction.@Rid as Attraction_RID, attraction.Name as Attraction_Name, attraction.Type as Attraction_Type' as RETURN
clause, this is the obtained list of records (only few records are shown in the image below):
Services
Example 1
Find all Services connected with Customer with OrderedId 1:
MATCH {class: Customers, as: customer, where: (OrderedId=1)}--{Class: Services, as: service}
RETURN $pathelements
In the Graph Editor included in Studio, using 'RETURN $pathelements' as RETURN
clause, this is the obtained graph:
In the Browse Tab of Studio, using 'RETURN service.@Rid as Service_RID, service.Name as Service_Name, service.Type as Service_Type' as RETURN
clause, this is the obtained list of records (only few records are shown in the image below):
Example 2
Find the 3 Hotels that have been booked most times:
SELECT
Name, Type, in("HasStayed").size() AS NumberOfBookings
FROM Hotels
ORDER BY NumberOfBookings DESC
LIMIT 3
In the Browse Tab of Studio, using the query below, this is the obtained list of records:
In a similar way:
Find the 3 Restaurants that have been used most times:
SELECT
Name, Type, in("HasEaten").size() AS VisitsNumber
FROM Restaurants
ORDER BY VisitsNumber DESC
LIMIT 3
In the Browse Tab of Studio, using the query below, this is the obtained list of records:
Example 3
Find the 3 Hotels that have most reviews:
SELECT
Name, Type, out("HasReview").size() AS ReviewNumbers
FROM `Hotels`
ORDER BY ReviewNumbers DESC
LIMIT 3
In the Browse Tab of Studio, using the query below, this is the obtained list of records:
In a similar way:
Find the 3 Restaurants that have most reviews :
SELECT
Name, Type, out("HasReview").size() AS ReviewNumbers
FROM `Restaurants`
ORDER BY ReviewNumbers DESC
LIMIT 3
In the Browse Tab of Studio, using the query below, this is the obtained list of records:
Example 4
Find the top 3 nationality of the tourists that have eaten at Restaurant with Id 26:
SELECT
Name,
count(*) as CountryCount
FROM (
SELECT
expand(out('IsFromCountry')) AS countries
FROM (
SELECT
expand(in("HasEaten")) AS customers
FROM Restaurants
WHERE Id='26'
UNWIND customers)
UNWIND countries)
GROUP BY Name
ORDER BY CountryCount DESC
LIMIT 3
In the Browse Tab of Studio, using the query below, this is the obtained list of records:
In a similar way:
Find the top 3 nationality of the tourists that stayed at Hotel with Id 627:
SELECT
Name, count(*) as CountryCount
FROM (
SELECT
expand(out('IsFromCountry')) AS countries
FROM (
SELECT
expand(in("HasStayed")) AS customers
FROM Hotels
WHERE Id='627'
UNWIND customers)
UNWIND countries)
GROUP BY Name
ORDER BY CountryCount DESC
LIMIT 3
In the Browse Tab of Studio, using the query below, this is the obtained list of records:
Locations
Example 1
Find all Locations visited by Santo:
MATCH {Class: Profiles, as: profile, where: (Name='Santo' AND Surname='OrientDB')}<-HasProfile-{Class: Customers, as: customer}-HasVisited->{class: Locations, as: location}
RETURN $pathelements
In the Graph Editor included in Studio, using the query above, this is the obtained graph:
Example 2
{% include "./include-file-7.md" %}
Example 3
{% include "./include-file-8.md" %}
Example 4
Find all Locations visited by Customer with OrderedId 2:
MATCH {Class: Locations, as: location}<-HasVisited-{class: Customers, as: customer, where: (OrderedId=2)}
RETURN $pathelements
In the Graph Editor included in Studio, using 'RETURN $pathelements' as RETURN
clause, this is the obtained graph:
In the Browse Tab of Studio, using 'RETURN location.@Rid as Location_RID, location.Name as Location_Name, location.Type as Location_Type' as RETURN
clause, this is the obtained list of records (only few records are shown in the image below):
Example 5
Find all Locations visited by Santo's friends:
MATCH {Class: Profiles, as: profile, where: (Name='Santo' and Surname='OrientDB')}-HasFriend->{Class: Profiles, as: friend}<-HasProfile-{Class: Customers, as: customer}-HasVisited->{Class: Locations, as: location}
RETURN $pathelements
In the Graph Editor included in Studio, using 'RETURN $pathelements' as RETURN
clause, this is the obtained graph:
In the Browse Tab of Studio, using 'location.@Rid as Location_RID, location.Name as Location_Name, location.Type as Location_Type, friend.Name as Friend_Name, friend.Surname as Friend_Surname' as RETURN
clause, this is the obtained list of records (only few records are shown in the image below):
Reviews
Example 1
Find number of Reviews per star:
SELECT
Stars, count(*) as Count
FROM HasReview
GROUP BY Stars
ORDER BY Count DESC
In the Browse Tab of Studio, using the query above, this is the obtained list of records:
Example 2
Find all reviewed Services:
MATCH {class: Services, as: s}-HasReview->{class: Reviews, as: r}
RETURN $pathelements
In the Graph Editor included in Studio, using the query above, this is the obtained graph:
Example 3
Find all reviewed Services and the Customer who made the review:
MATCH {class: Services, as: s}-HasReview->{class: Reviews, as: r}<-MadeReview-{class: Customers, as: c}
RETURN $pathelements
In the Graph Editor included in Studio, using the query above, this is the obtained graph:
Example 4
Find the numbers of reviews per Service:
SELECT
@rid as Service_RID,
Name as Service_Name,
Type as Service_Type,
out("HasReview").size() AS ReviewNumbers
FROM `Services`
ORDER BY ReviewNumbers DESC
In the Browse Tab of Studio, using the query above, this is the obtained list of records:
Example 5
{% include "./include-file-2.md" %}
Example 6
{% include "./include-file-3.md" %}
Recommendations
Example 1
Recommend some friends to Profile 'Isabella Gomez' (friends of friends):
MATCH
{class: Profiles, as: profile, where: (Name = 'Isabella' AND Surname='Gomez')}-HasFriend-{as: friend},
{as: friend}-HasFriend-{as: friendOfFriend, where: ($matched.profile not in $currentMatch.both('HasFriend') and $matched.profile != $currentMatch)}
RETURN DISTINCT friendOfFriend.Name
In the Browse Tab of Studio, using the query above, this is the obtained list of records (only few records are shown in the image below):
Example 2
Recommend some Hotels to Customer with OrderedId 1:
MATCH
{Class: Customers, as: customer, where: (OrderedId=1)}-HasProfile->{class: Profiles, as: profile},
{as: profile}-HasFriend->{class: Profiles, as: friend},
{as: friend}<-HasProfile-{Class: Customers, as: customerFriend},
{as: customerFriend}-HasStayed->{Class: Hotels, as: hotel},
{as: customerFriend}-MadeReview->{Class: Reviews, as: review},
{as: hotel}-HasReview->{as: review}
RETURN $pathelements
In the Graph Editor included in Studio, using 'RETURN $pathelements' as RETURN
clause, this is the obtained graph:
To filter additionally, and suggest only the 4 and 5-rated hotels, it is possible to add a filter condition on the 'HasReview' edge (property 'Stars'):
MATCH
{Class: Customers, as: customer, where: (OrderedId=1)}-HasProfile->{class: Profiles, as: profile},
{as: profile}-HasFriend->{class: Profiles, as: friend},
{as: friend}<-HasProfile-{Class: Customers, as: customerFriend},
{as: customerFriend}-HasStayed->{Class: Hotels, as: hotel},
{as: customerFriend}-MadeReview->{Class: Reviews, as: review},
{as: hotel}.outE('HasReview'){as: ReviewStars, where: (Stars>3)}.inV(){as: review}
RETURN $pathelements
In the Graph Editor included in Studio, using 'RETURN $pathelements' as RETURN
clause, this is the obtained graph:
Business Opportunities
Example 1
{% include "./include-file-1.md" %}
Example 2
Find all the Customer Friends that are not Customers (so that a product can be proposed).
In the Graph Editor included in Studio, using the query below, this is the obtained graph:
SELECT DISTINCT * FROM (
SELECT expand(customerFriend)
FROM (
MATCH
{Class:Customers, as: customer}-HasProfile-{Class:Profiles, as: profile}-HasFriend-{Class:Profiles, as: customerFriend}
RETURN customerFriend
)
)
WHERE in('HasProfile').size()=0
In the Browse Tab of Studio, using the query below, this is the obtained list of records (only few records are shown in the image below):
SELECT DISTINCT @Rid as Friend_RID, Name as Friend_Name, Surname as Friend_Surname
FROM (
SELECT expand(customerFriend)
FROM (
MATCH
{Class:Customers, as: customer}-HasProfile-{Class:Profiles, as: profile}-HasFriend-{Class:Profiles, as: customerFriend}
RETURN customerFriend
)
)
WHERE in('HasProfile').size()=0
ORDER BY Friend_RID
Polymorphism
Example 1
Find all Locations (Services + Attractions) connected with Customer with OrderedId 1:
MATCH {class: Customers, as: customer, where: (OrderedId=1)}--{Class: Locations, as: location}
RETURN $pathelements
In the Graph Editor included in Studio, using 'RETURN $pathelements' as RETURN
clause, this is the obtained graph:
In the Browse Tab of Studio, using 'RETURN location.@Rid as Location_RID, location.Name as Location_Name, location.Type as Location_Type' as RETURN
clause, this is the obtained list of records (only few records are shown in the image below):
Example 2
{% include "./include-file-2.md" %}
Shortest Paths
Example 1
Find the shortest path between the Profile 'Santo' and the Country 'United States':
SELECT expand(path) FROM (
SELECT shortestPath($from, $to) AS path
LET
$from = (SELECT FROM Profiles WHERE Name='Santo' and Surname='OrientDB'),
$to = (SELECT FROM Countries WHERE Name='United States')
UNWIND path
)
In the Graph Editor included in Studio, this is the obtained graph:
In the Browse Tab of Studio, this is the obtained list of records:
Example 2
Find the shortest path between the Profile 'Santo' and the Restaurant 'Malga Granezza':
SELECT expand(path) FROM (
SELECT shortestPath($from, $to) AS path
LET
$from = (SELECT FROM Profiles WHERE Name='Santo' and Surname='OrientDB'),
$to = (SELECT FROM Restaurants WHERE Name='Malga Granezza')
UNWIND path
)
In the Graph Editor included in Studio, this is the obtained graph:
In the Browse Tab of Studio, this is the obtained list of records:
Traverses
Example 1
Traverse everything from Profile 'Santo' up to depth three:
TRAVERSE * FROM (
SELECT FROM Profiles WHERE Name='Santo' and Surname='OrientDB'
) MAXDEPTH 3
In the Graph Editor included in Studio, this is the obtained graph:
In the Browse Tab of Studio, this is the obtained list of records (only few records are shown in the image below):
Example 2
Traverse everything from Country 'Italy' up to depth three:
TRAVERSE * FROM (
SELECT FROM Countries WHERE Name='Italy'
) MAXDEPTH 3
In the Graph Editor included in Studio, this is the obtained graph:
In the Browse Tab of Studio, this is the obtained list of records (only few records are shown in the image below):
Graph Portions
This Section includes some screen-shots on "portions" of the demodb
graph.
{% include "./include-demodb-version-warning.md" %}
Profiles
The graph below includes vertices in the Profiles class.
Some profiles have no friends, and ten profile "communities" are easily identifiable (a "community" is a group of highly connected profiles):
Customers' Friends
The graph below shows the relationships between all customers and their friends.
All customer vertices are connected to their corresponding profiles in the social platform. The profiles may have friends, or not.
Customer-Country Relationship
The graph below shows the relationships between all customers and their countries of origin:
Customer-Restaurant Relationship
Relationships between customers and restaurants are reported in the graph below:
SQL
Most NoSQL products employ a custom query language. In this, OrientDB differs by focusing on standards in query languages. That is, instead of inventing "Yet Another Query Language," it begins with the widely used and well-understood language of SQL. It then extends SQL to support more complex graphing concepts, such as Trees and Graphs.
Why SQL? Because SQL is ubiquitous in the database development world. It is familiar and more readable and concise than its competitors, such as Map Reduce scripts or JSON based querying.
SELECT
The SELECT
statement queries the database and returns results that match the given parameters. For instance, earlier in Getting Started, two queries were presented that gave the same results: BROWSE CLUSTER ouser
and BROWSE CLASS OUser
. Here is a third option, available through a SELECT
statement.
orientdb> SELECT FROM OUser
Notice that the query has no projections. This means that you do not need to enter a character to indicate that the query should return the entire record, such as the asterisk in the Relational model, (that is, SELECT * FROM OUser
).
Additionally, OUser is a class. By default, OrientDB executes queries against classes. Targets can also be:
-
Clusters To execute against a cluster, rather than a class, prefix
CLUSTER
to the target name.orientdb>
SELECT FROM CLUSTER:Ouser
-
Record ID To execute against one or more Record ID's, use the identifier(s) as your target. For example.
orientdb>
SELECT FROM #10:3
orientdb>SELECT FROM [#10:1, #10:30, #10:5]
-
Indexes To execute a query against an index, prefix
INDEX
to the target name.orientdb>
SELECT VALUE FROM INDEX:dictionary WHERE key='Jay'
WHERE
Much like the standard implementation of SQL, OrientDB supports WHERE
conditions to filter the returning records too. For example,
orientdb> SELECT FROM OUser WHERE name LIKE 'l%'
This returns all OUser
records where the name begins with l
. For more information on supported operators and functions, see WHERE
.
ORDER BY
In addition to WHERE
, OrientDB also supports ORDER BY
clauses. This allows you to order the results returned by the query according to one or more fields, in either ascending or descending order.
orientdb> SELECT FROM Employee WHERE city='Rome' ORDER BY surname ASC, name ASC
The example queries the Employee
class, it returns a listing of all employees in that class who live in Rome and it orders the results by surname and name, in ascending order.
GROUP BY
In the event that you need results of the query grouped together according to the values of certain fields, you can manage this using the GROUP BY
clause.
orientdb> SELECT SUM(salary) FROM Employee WHERE age < 40 GROUP BY job
In the example, you query the Employee
class for the sum of the salaries of all employees under the age of forty, grouped by their job types.
LIMIT
In the event that your query returns too many results, making it difficult to read or manage, you can use the LIMIT
clause to reduce it to the top most of the return values.
orientdb> SELECT FROM Employee WHERE gender='male' LIMIT 20
In the example, you query the Employee
class for a list of male employees. Given that there are likely to be a number of these, you limit the return to the first twenty entries.
SKIP
When using the LIMIT
clause with queries, you can only view the topmost of the return results. In the event that you would like to view certain results further down the list, for instance the values from twenty to forty, you can paginate your results using the SKIP
keyword in the LIMIT
clause.
orientdb>SELECT FROM Employee WHERE gender='male' LIMIT 20
orientdb>SELECT FROM Employee WHERE gender='male' SKIP 20 LIMIT 20
orientdb>SELECT FROM Employee WHERE gender='male' SKIP 40 LIMIT 20
The first query returns the first twenty results, the second returns the next twenty results, the third up to sixty. You can use these queries to manage pages at the application layer.
INSERT
The INSERT
statement adds new data to a class and cluster. OrientDB supports three forms of syntax used to insert new data into your database.
-
The standard ANSI-92 syntax:
orientdb>
INSERT INTO Employee(name, surname, gender) VALUES('Jay', 'Miner', 'M')
-
The simplified ANSI-92 syntax:
orientdb>
INSERT INTO Employee SET name='Jay', surname='Miner', gender='M'
-
The JSON syntax:
orientdb>
INSERT INTO Employee CONTENT
{name : 'Jay', surname : 'Miner', gender : 'M'}
Each of these queries adds Jay Miner to the Employee
class. You can choose whichever syntax that works best with your application.
UPDATE
The UPDATE
statement changes the values of existing data in a class and cluster. In OrientDB there are two forms of syntax used to update data on your database.
-
The standard ANSI-92 syntax:
orientdb>
UPDATE Employee SET local=TRUE WHERE city='London'
-
The JSON syntax, used with the
MERGE
keyword, which merges the changes with the current record:orientdb>
UPDATE Employee MERGE { local : TRUE } WHERE city='London'
Each of these statements updates the Employee
class, changing the local
property to TRUE
when the employee is based in London.
DELETE
The DELETE
statement removes existing values from your class and cluster. OrientDB supports the standard ANSI-92 compliant syntax for these statements:
orientdb> DELETE FROM Employee WHERE city <> 'London'
Here, entries are removed from the Employee
class where the employee in question is not based in London.
See also:
Users, Roles and Security
Distributed Architecture
search: keywords: ['tutorials']
Tutorials
This Section includes all the Tutorials available in this Manual.
Working with Graphs
In graph databases, the database system graphs data into network-like structures consisting of vertices and edges. In the OrientDB Graph model, the database represents data through the concept of a property graph, which defines a vertex as an entity linked with other vertices and an edge, as an entity that links two vertices.
OrientDB ships with a generic vertex persistent class, called V
, as well as a class for edges, called E
. As an example, you can create a new vertex using the INSERT
command with V
.
orientdb> INSERT INTO V SET name='Jay'
Created record with RID #9:0
In effect, the Graph model database works on top of the underlying document model. But, in order to simplify this process, OrientDB introduces a new set of commands for managing graphs from the console. Instead of INSERT
, use CREATE VERTEX
orientdb> CREATE VERTEX V SET name='Jay'
Created vertex with RID #9:1
By using the graph commands over the standard SQL syntax, OrientDB ensures that your graphs remain consistent. For more information on the particular commands, see the following pages:
Use Case: Social Network for Restaurant Patrons
While you have the option of working with vertexes and edges in your database as they are, you can also extend the standard V
and E
classes to suit the particular needs of your application. The advantages of this approach are,
- It grants better understanding about the meaning of these entities.
- It allows for optional constraints at the class level.
- It improves performance through better partitioning of entities.
- It allows for object-oriented inheritance among the graph elements.
For example, consider a social network based on restaurants. You need to start with a class for individual customers and another for the restaurants they patronize. Create these classes to extend the V
class.
orientdb>CREATE CLASS Person EXTENDS V
orientdb>CREATE CLASS Restaurant EXTENDS V
Doing this creates the schema for your social network. Now that the schema is ready, populate the graph with data.
orientdb>CREATE VERTEX Person SET name='Luca'
Created record with RID #11:0 orientdb>CREATE VERTEX Person SET name='Bill'
Created record with RID #11:1 orientdb>CREATE VERTEX Person SET name='Jay'
Created record with RID #11:2 orientdb>CREATE VERTEX Restaurant SET name='Dante', type='Pizza'
Created record with RID #12:0 orientdb>CREATE VERTEX Restaurant SET name='Charlie', type='French'
Created record with RID #12:1
This adds three vertices to the Person
class, representing individual users in the social network. It also adds two vertices to the Restaurant
class, representing the restaurants that they patronize.
Creating Edges
For the moment, these vertices are independent of one another, tied together only by the classes to which they belong. That is, they are not yet connected by edges. Before you can make these connections, you first need to create a class that extends E
.
orientdb> CREATE CLASS Eat EXTENDS E
This creates the class Eat
, which extends the class E
. Eat
represents the relationship between the vertex Person
and the vertex Restaurant
.
When you create the edge from this class, note that the orientation of the vertices is important, because it gives the relationship its meaning. For instance, creating an edge in the opposite direction, (from Restaurant
to Person
), would call for a separate class, such as Attendee
.
The user Luca eats at the pizza joint Dante. Create an edge that represents this connection:
orientdb> CREATE EDGE Eat FROM ( SELECT FROM Person WHERE name='Luca' )
TO ( SELECT FROM Restaurant WHERE name='Dante' )
Creating Edges from Record ID
In the event that you know the Record ID of the vertices, you can connect them directly with a shorter and faster command. For example, the person Bill also eats at the restaurant Dante and the person Jay eats at the restaurant Charlie. Create edges in the class Eat
to represent these connections.
orientdb>CREATE EDGE Eat FROM #11:1 TO #12:0
orientdb>CREATE EDGE Eat FROM #11:2 TO #12:1
Querying Graphs
In the above example you created and populated a small graph of a social network of individual users and the restaurants at which they eat. You can now begin to experiment with queries on a graph database.
To cross edges, you can use special graph functions, such as:
OUT()
To retrieve the adjacent outgoing verticesIN()
To retrieve the adjacent incoming verticesBOTH()
To retrieve the adjacent incoming and outgoing vertices
For example, to know all of the people who eat in the restaurant Dante, which has a Record ID of #12:0
, you can access the record for that restaurant and traverse the incoming edges to discover which entries in the Person
class connect to it.
orientdb> SELECT IN() FROM Restaurant WHERE name='Dante'
-------+----------------+
@RID | in |
-------+----------------+
#-2:1 | [#11:0, #11:1] |
-------+----------------+
This query displays the record ID's from the Person
class that connect to the restaurant Dante. In cases such as this, you can use the EXPAND()
special function to transform the vertex collection in the result-set by expanding it.
orientdb> SELECT EXPAND( IN() ) FROM Restaurant WHERE name='Dante'
-------+-------------+-------------+---------+
@RID | @CLASS | Name | out_Eat |
-------+-------------+-------------+---------+
#11:0 | Person | Luca | #12:0 |
#11:1 | Person | Bill | #12:0 |
-------+-------------+-------------+---------+
Creating Edge to Connect Users
Your application at this point shows connections between individual users and the restaurants they patronize. While this is interesting, it does not yet function as a social network. To do so, you need to establish edges that connect the users to one another.
To begin, as before, create a new class that extends E
:
orientdb> CREATE CLASS Friend EXTENDS E
The users Luca and Jay are friends. They have Record ID's of #11:0
and #11:2
. Create an edge that connects them.
orientdb> CREATE EDGE Friend FROM #11:0 TO #11:2
In the Friend
relationship, orientation is not important. That is, if Luca is a friend of Jay's then Jay is a friend of Luca's. Therefore, you should use the BOTH()
function.
orientdb> SELECT EXPAND( BOTH( 'Friend' ) ) FROM Person WHERE name = 'Luca'
-------+-------------+-------------+---------+-----------+
@RID | @CLASS | Name | out_Eat | in_Friend |
-------+-------------+-------------+---------+-----------+
#11:2 | Person | Jay | #12:1 | #11:0 |
-------+-------------+-------------+---------+-----------+
Here, the BOTH()
function takes the edge class Friend
as an argument, crossing only relationships of the Friend kind, (that is, it skips the Eat
class, at this time). Note in the result-set that the relationship with Luca, with a Record ID of #11:0
in the in_
field.
You can also now view all the restaurants patronized by friends of Luca.
orientdb> SELECT EXPAND( BOTH('Friend').out('Eat') ) FROM Person
WHERE name='Luca'
-------+-------------+-------------+-------------+--------+
@RID | @CLASS | Name | Type | in_Eat |
-------+-------------+-------------+-------------+--------+
#12:1 | Restaurant | Charlie | French | #11:2 |
-------+-------------+-------------+-------------+--------+
Lightweight Edges
In version 1.4.x, OrientDB begins to manage some edges as Lightweight Edges. Lightweight Edges do not have Record ID's, but are physically stored as links within vertices. Note that OrientDB only uses a Lightweight Edge only when the edge has no properties, otherwise it uses the standard Edge.
From the logic point of view, Lightweight Edges are Edges in all effects, so that all graph functions work with them. This is to improve performance and reduce disk space.
Because Lightweight Edges don't exist as separate records in the database, some queries won't work as expected. For instance,
orientdb> SELECT FROM E
For most cases, an edge is used connecting vertices, so this query would not cause any problems in particular. But, it would not return Lightweight Edges in the result-set. In the event that you need to query edges directly, including those with no properties, disable the Lightweight Edge feature.
To disable the Lightweight Edge feature, execute the following command.
orientdb> ALTER DATABASE CUSTOM useLightweightEdges=FALSE
You only need to execute this command once. OrientDB now generates new edges as the standard Edge, rather than the Lightweight Edge. Note that this does not affect existing edges.
For troubleshooting information on Lightweight Edges, see Why I can't see all the edges. For more information in the Graph model in OrientDB, see Graph API.
Using Schema with Graphs
OrientDB, through the Graph API, offers a number of features above and beyond the traditional Graph Databases given that it supports concepts drawn from both the Document Database and the Object Oriented worlds. For instance, consider the power of graphs, when used in conjunction with schemas and constraints.
Use Case: Car Database
For this example, consider a graph database that maps the relationship between individual users and their cars. First, create the graph schema for the Person
and Car
vertex classes, as well as the Owns
edge class to connect the two:
orientdb>CREATE CLASS Person EXTENDS V
orientdb>CREATE CLASS Car EXTENDS V
orientdb>CREATE CLASS Owns EXTENDS E
These commands lay out the schema for your graph database. That is, they define two vertex classes and an edge class to indicate the relationship between the two. With that, you can begin to populate the database with vertices and edges.
orientdb>CREATE VERTEX Person SET name = 'Luca'
Created vertex 'Person#11:0{name:Luca} v1' in 0,012000 sec(s). orientdb>CREATE VERTEX Car SET name = 'Ferrari Modena'
Created vertex 'Car#12:0{name:Ferrari Modena} v1' in 0,001000 sec(s). orientdb>CREATE EDGE Owns FROM ( SELECT FROM Person ) TO ( SELECT FROM Car )
Created edge '[e[#11:0->#12:0][#11:0-Owns->#12:0]]' in 0,005000 sec(s).
Querying the Car Database
In the above section, you create a car database and populated it with vertices and edges to map out the relationship between drivers and their cars. Now you can begin to query this database, showing what those connections are. For example, what is Luca's car? You can find out by traversing from the vertex Luca to the outgoing vertices following the Owns
relationship.
orientdb> SELECT name FROM ( SELECT EXPAND( OUT('Owns') ) FROM Person
WHERE name='Luca' )
----+-------+-----------------+
# | @RID | name |
----+-------+-----------------+
0 | #-2:1 | Ferrari Modena |
----+-------+-----------------+
As you can see, the query returns that Luca owns a Ferrari Modena. Now consider expanding your database to track where each person lives.
Adding a Location Vertex
Consider a situation, in which you might want to keep track of the countries in which each person lives. In practice, there are a number of reasons why you might want to do this, for instance, for the purposes of promotional material or in a larger database to analyze the connections to see how residence affects car ownership.
To begin, create a vertex class for the country, in which the person lives and an edge class that connects the individual to the place.
orientdb>CREATE CLASS Country EXTENDS V
orientdb>CREATE CLASS Lives EXTENDS E
This creates the schema for the feature you're adding to the cars database. The vertex class Country
recording countries in which people live and the edge class Lives
to connect individuals in the vertex class Person
to entries in Country
.
With the schema laid out, create a vertex for the United Kingdom and connect it to the person Luca.
orientdb>CREATE VERTEX Country SET name='UK'
Created vertex 'Country#14:0{name:UK} v1' in 0,004000 sec(s). orientdb>CREATE EDGE Lives FROM ( SELECT FROM Person ) TO ( SELECT FROM Country
Created edge '[e[#11:0->#14:0][#11:0-Lives->#14:0]]' in 0,006000 sec(s).
The second command creates an edge connecting the person Luca to the country United Kingdom. Now that your cars database is defined and populated, you can query it, such as a search that shows the countries where there are users that own a Ferrari.
orientdb> SELECT name FROM ( SELECT EXPAND( IN('Owns').OUT('Lives') )
FROM Car WHERE name LIKE '%Ferrari%' )
---+-------+--------+
# | @RID | name |
---+-------+--------+
0 | #-2:1 | UK |
---+-------+--------+
Using in
and out
Constraints on Edges
In the above sections, you modeled the graph using a schema without any constraints, but you might find it useful to use some. For instance, it would be good to require that an Owns
relationship only exist between the vertex Person
and the vertex Car
.
orientdb>CREATE PROPERTY Owns.out LINK Person
orientdb>CREATE PROPERTY Owns.in LINK Car
These commands link outgoing vertices of the Person
class to incoming vertices of the Car
class. That is, it configures your database so that a user can own a car, but a car cannot own a user.
Using MANDATORY
Constraints on Edges
By default, when OrientDB creates an edge that lacks properties, it creates it as a Lightweight Edge. That is, it creates an edge that has no physical record in the database. Using the MANDATORY
setting, you can stop this behavior, forcing it to create the standard Edge, without outright disabling Lightweight Edges.
orientdb>ALTER PROPERTY Owns.out MANDATORY TRUE
orientdb>ALTER PROPERTY Owns.in MANDATORY TRUE
Using UNIQUE
with Edges
For the sake of simplicity, consider a case where you want to limit the way people are connected to cars to where the user can only match to the car once. That is, if Luca owns a Ferrari Modena, you might prefer not to have a double entry for that car in the event that he buys a new one a few years later. This is particularly important given that our database covers make and model, but not year.
To manage this, you need to define a UNIQUE
index against both the out and in properties.
orientdb> CREATE INDEX UniqueOwns ON Owns(out,in) UNIQUE
Created index successfully with 0 entries in 0,023000 sec(s).
The index returns tells us that no entries are indexed. You have already created the Owns
relationship between Luca and the Ferrari Modena. In that case, however, OrientDB had created a Lightweight Edge before you set the rule to force the creation of documents for Owns
instances. To fix this, you need to drop and recreate the edge.
orientdb>DELETE EDGE FROM #11:0 TO #12:0
orientdb>CREATE EDGE Owns FROM ( SELECT FROM Person ) TO ( SELECT FROM Car )
To confirm that this was successful, run a query to check that a record was created:
orientdb> SELECT FROM Owns
---+-------+-------+--------+
# | @RID | out | in |
---+-------+-------+--------+
0 | #13:0 | #11:0 | #12:0 |
---+-------+-------+--------+
This shows that a record was indeed created. To confirm that the constraints work, attempt to create an edge in Owns
that connects Luca to the United Kingdom.
orientdb> CREATE EDGE Owns FROM ( SELECT FROM Person ) TO ( SELECT FROM Country )
Error: com.orientechnologies.orient.core.exception.OCommandExecutionException:
Error on execution of command: sql.create edge Owns from (select from Person)...
Error: com.orientechnologies.orient.core.exception.OValidationException: The
field 'Owns.in' has been declared as LINK of type 'Car' but the value is the
document #14:0 of class 'Country'
This shows that the constraints effectively blocked the creation, generating a set of errors to explain why it was blocked.
You now have a typed graph with constraints. For more information, see Graph Schema.
Setting up a Distributed Graph Database
In addition to the standard deployment architecture, where it runs as a single, standalone database instance, you can also deploy OrientDB using Distributed Architecture. In this environment, it shares the database across multiple server instances.
Launching Distributed Server Cluster
There are two ways to share a database across multiple server nodes:
-
Prior to startup, copy the specific database directory, under
$ORIENTDB_HOME/database
to all servers. -
Keep the database on the first running server node, then start every other server node. Under the default configurations, OrientDB automatically shares the database with the new servers that join.
This tutorial assumes that you want to start a distributed database using the second method.
NOTE: When you run in distributed mode, OrientDB needs more RAM. The minimum is 2GB of heap, but we suggest to use at least 4GB of heap memory. To change the heap modify the Java memory settings in the file bin/dserver.sh
(or dserver.bat on Windows).
Starting the First Server Node
Unlike the standard standalone deployment of OrientDB, there is a different script that you need to use when launching a distributed server instance. Instead of server.sh
, you use dserver.sh
. In the case of Windows, use dserver.bat
. Whichever you need, you can find it in the bin
of your installation directory.
$ ./bin/dserver.sh
Bear in mind that OrientDB uses the same orientdb-server-config.xml
configuration file, regardless of whether it's running as a server or distributed server. For more information, see Distributed Configuration.
The first time you start OrientDB as a distributed server, it generates the following output:
+---------------------------------------------------------------+
| WARNING: FIRST DISTRIBUTED RUN CONFIGURATION |
+---------------------------------------------------------------+
| This is the first time that the server is running as |
| distributed. Please type the name you want to assign to the |
| current server node. |
| |
| To avoid this message set the environment variable or JVM |
| setting ORIENTDB_NODE_NAME to the server node name to use. |
+---------------------------------------------------------------+
Node name [BLANK=auto generate it]:
You need to give the node a name here. OrientDB stores it in the nodeName
parameter of OHazelcastPlugin
. It adds the variable to your orientdb-server-config.xml
configuration file.
Distributed Startup Process
When OrientDB starts as a distributed server instance, it loads all databases in the database
directory and configures them to run in distributed mode. For this reason, the first load, OrientDB copies the default distributed configuration, (that is, the default-distributed-db-config.json
configuration file), into each database's directory, renaming it distributed-config.json
. On subsequent starts, each database uses this file instead of the default configuration file. Since the shape of the cluster changes every time nodes join or leave, the configuration is kept up to date by each distributed server instance.
For more information on working with the default-distributed-db-config.json
configuration file, see Distributed Configuration.
Starting Additional Server Nodes
When you have the first server node running, you can begin to start the other server nodes. Each server requires the same Hazelcast credentials in order to join the same cluster. You can define these in the hazelcast.xml
configuration file.
The fastest way to initialize multiple server nodes is to copy the OrientDB installation directory from the first node to each of the subsequent nodes. For instance,
$ scp user@ip_address $ORIENTDB_HOME
This copies both the databases and their configuration files onto the new distributed server node.
Bear in mind, if you run multiple server instances on the same host, such as when testing, you need to change the port entry in the
hazelcast.xml
configuration file.
For the other server nodes in the cluster, use the same dserver.sh
command as you used in starting the first node. When the other server nodes come online, they begin to establish network connectivity with each other. Monitoring the logs, you can see where they establish connections from messages such as this:
WARN [node1384014656983] added new node id=Member [192.168.1.179]:2435 name=null
[OHazelcastPlugin]
INFO [192.168.1.179]:2434 [orientdb] Re-partitioning cluster data... Migration
queue size: 135 [PartitionService]
INFO [192.168.1.179]:2434 [orientdb] All migration tasks has been completed,
queues are empty. [PartitionService]
INFO [node1384014656983] added node configuration id=Member [192.168.1.179]:2435
name=node1384015873680, now 2 nodes are configured [OHazelcastPlugin]
INFO [node1384014656983] update configuration db=GratefulDeadConcerts
from=node1384015873680 [OHazelcastPlugin]
INFO updated distributed configuration for database: GratefulDeadConcerts:
----------
{
"replication": true,
"autoDeploy": true,
"hotAlignment": true,
"resyncEvery": 15,
"clusters": {
"internal": {
"replication": false
},
"index": {
"replication": false
},
"*": {
"replication": true,
"readQuorum": 1,
"writeQuorum": 2,
"failureAvailableNodesLessQuorum": false,
"readYourWrites": true,
"partitioning":{
"strategy": "round-robin",
"default":0,
"partitions": ["","node1383734730415","node1384015873680"]("","node1383734730415","node1384015873680".md)
}
}
},
"version": 1
}
---------- [OHazelcastPlugin]
WARN [node1383734730415]->[node1384015873680] deploying database
GratefulDeadConcerts...[ODeployDatabaseTask]
WARN [node1383734730415]->[node1384015873680] sending the compressed database
GratefulDeadConcerts over the network, total 339,66Kb [ODeployDatabaseTask]
In the example, two server nodes were started on the same machine. It has an IP address of 10.37.129.2, but is using OrientDB on two different ports: 2434 and 2435, where the current is called this
. The remainder of the log is relative to the distribution of the database to the second server.
On the second server node output, OrientDB dumps messages like this:
WARN [node1384015873680]<-[node1383734730415] installing database
GratefulDeadConcerts in databases/GratefulDeadConcerts... [OHazelcastPlugin]
WARN [node1384015873680] installed database GratefulDeadConcerts in
databases/GratefulDeadConcerts, setting it online... [OHazelcastPlugin]
WARN [node1384015873680] database GratefulDeadConcerts is online [OHazelcastPlugin]
WARN [node1384015873680] updated node status to 'ONLINE' [OHazelcastPlugin]
INFO OrientDB Server v1.6.1-SNAPSHOT is active. [OServer]
What these messages mean is that the database GratefulDeadConcerts
was correctly installed from the first node, that is node1383734730415
through the network.
Migrating from standalone server to a cluster
If you have a standalone instance of OrientDB and you want to move to a cluster you should follow these steps:
- Install OrientDB on all the servers of the cluster and configure it (according to the sections above)
- Stop the standalone server
- Copy the specific database directories under
$ORIENTDB_HOME/database
to all the servers of the cluster - Start all the servers in the cluster using the script
dserver.sh
(ordserver.bat
if on Windows)
If the standalone server will be part of the cluster, you can use the existing installation of OrientDB; you don't need to copy the database directories since they're already in place and you just have to start it before all the other servers with dserver.sh
.
Working with Distributed Graphs
When OrientDB joins a distributed cluster, all clients connecting to the server node are constantly notified about this state. This ensures that, in the event that server node fails, the clients can switch transparently to the next available server.
You can check this through the console. When OrientDB runs in a distributed configuration, the current cluster shape is visible through the INFO
command.
$$ORIENTDB_HOME/bin/console.sh
OrientDB console v.1.6 www.orientechnologies.com Type 'help' to display all the commands supported. Installing extensions for GREMLIN language v.2.5.0-SNAPSHOT orientdb>CONNECT remote:localhost/GratefulDeadConcerts admin admin
Connecting to database [remote:localhost/GratefulDeadConcerts] with user 'admin'...OK orientdb>INFO
Current database: GratefulDeadConcerts (url=remote:localhost/GratefulDeadConcerts)
For reference purposes, the server nodes in the example have the following configurations. As you can see, it is a two node cluster running a single server host. The first node listens on port 2481
while the second on port 2480
.
+---------+------+-----------------------------------------+-----+---------+--------------+--------------+-----------------------+
|Name |Status|Databases |Conns|StartedOn|Binary |HTTP |UsedMemory |
+---------+------+-----------------------------------------+-----+---------+--------------+--------------+-----------------------+
|europe-0 |ONLINE|distributed-node-deadlock=ONLINE (MASTER)|5 |16:53:59 |127.0.0.1:2424|127.0.0.1:2480|269.32MB/3.56GB (7.40%)|
|europe-1 |ONLINE|distributed-node-deadlock=ONLINE (MASTER)|4 |16:54:03 |127.0.0.1:2425|127.0.0.1:2481|268.89MB/3.56GB (7.38%)|
+---------+------+-----------------------------------------+-----+---------+--------------+--------------+-----------------------+
Testing Distributed Architecture
Once you have a distributed database up and running, you can begin to test its operations on a running environment. For example, begin by creating a vertex, setting the node
property to 1
.
orientdb> CREATE VERTEX V SET node = 1
Created vertex 'V#9:815{node:1} v1' in 0,013000 sec(s).
From another console, connect to the second node and execute the following command:
orientdb> SELECT FROM V WHERE node = 1
----+--------+-------+
# | @RID | node |
----+--------+-------+
0 | #9:815 | 1 |
----+--------+-------+
1 item(s) found. Query executed in 0.19 sec(s).
This shows that the vertex created on the first node has successfully replicated to the second node.
Logs in Distributed Architecture
From time to time server nodes go down. This does not necessarily relate to problems in OrientDB, (for instance, it could originate from limitations in system resources).
To test this out, kill the first node. For example, assuming the first node has a process identifier, (that is, a PID), of 1254
on your system, run the following command:
$ kill -9 1254
This command kills the process on PID 1254
. Now, check the log messages for the second node:
$less orientdb.log
INFO [127.0.0.1]:2435 [orientdb] Removing Member [127.0.0.1]:2434 [ClusterService] INFO [127.0.0.1]:2435 [orientdb]Members [1] { Member [127.0.0.1]:2435 this }
[ClusterService] WARN [europe-0] node removed id=Member [127.0.0.1]:2434 name=europe-1 [OHazelcastPlugin] INFO [127.0.0.1]:2435 [orientdb] Partition balance is ok, no need to re-partition cluster data... [PartitionService]
What the logs show you is that the second node is now aware that it cannot reach the first node. You can further test this by running the console connected to the first node..
orientdb> SELECT FROM V LIMIT 2
WARN Caught I/O errors from /127.0.0.1:2425 (local
socket=0.0.0.0/0.0.0.0:51512), trying to reconnect (error:
java.io.IOException: Stream closed) [OStorageRemote]
WARN Connection re-acquired transparently after 30ms and 1 retries: no errors
will be thrown at application level [OStorageRemote]
---+------+----------------+--------+--------------+------+-----------------+-----
# | @RID | name | song_type | performances | type | out_followed_by | ...
---+------+----------------+--------+--------------+------+-----------------+-----
1 | #9:1 | HEY BO DIDDLEY | cover | 5 | song | [5] | ...
2 | #9:2 | IM A MAN | cover | 1 | song | [2] | ...
---+------+----------------+--------+--------------+------+-----------------+-----
This shows that the console auto-switched to the next available node. That is, it switched to the second node upon noticing that the first was no longer functional. The warnings reports show what happened in a transparent way, so that the application doesn't need to manage the issue.
From the console connected to the second node, create a new vertex.
orientdb> CREATE VERTEX V SET node=2
Created vertex 'V#9:816{node:2} v1' in 0,014000 sec(s).
Given that the first node remains nonfunctional, OrientDB journals the operation. Once the first node comes back online, the second node synchronizes the changes into it.
Restart the first node and check that it successfully auto-realigns. Reconnect the console to the first node and run the following command:
orientdb> SELECT FROM V WHERE node=2
---+--------+-------+
# | @RID | node |
---+--------+-------+
0 | #9:816 | 2 |
---+--------+-------+
1 item(s) found. Query executed in 0.209 sec(s).
This shows that the first node has realigned itself with the second node.
This process is repeatable with N server nodes, where every server is a master. There is no limit to the number of running servers. With many servers spread across a slow network, you can tune the network timeouts to be more permissive and let a large, distributed cluster of servers work properly.
For more information, Distributed Architecture.
Time Series Use Case
Managing records related to historical information is pretty common. When you have millions of records, indexes start show their limitations, because the cost to find the records is O(logN). This is also the main reason why Relational DBMS are so slow with huge databases.
So when you have millions of record the best way to scale up linearly is avoid using indexes at all or as much as you can. But how can you retrieve records in a short time without indexes? Should OrientDB scan the entire database at every query? No. You should use the Graph properties of OrientDB. Let's look at a simple example, where the domain are logs.
A typical log record has some information about the event and a date. Below is the Log record to use in our example. We're going to use the JSON format to simplify reading:
{
"date" : 12293289328932,
"priority" : "critical",
"note" : "System reboot"
}
Now let's create a tree (that is a directed, non cyclic graph) to group the Log records based on the granularity we need. Example:
Year -> month (map) -> Month -> day (map) -> Day -> hour (map) -> Hour
Where Year, Month, Day and Hour are vertex classes. Each Vertex links the other Vertices of smaller type. The links should be handled using a Map to make easier the writing of queries.
Create the classes:
CREATE CLASS Year
CREATE CLASS Month
CREATE CLASS Day
CREATE CLASS Hour
CREATE PROPERTY Year.month LINKMAP Month
CREATE PROPERTY Month.day LINKMAP Day
CREATE PROPERTY Day.hour LINKMAP Hour
Example to retrieve the vertex relative to the date March 2012, 20th at 10am (2012/03/20 10:00:00):
SELECT month[3].day[20].hour[10].logs FROM Year WHERE year = "2012"
If you need more granularity than the Hour you can go ahead until the Time unit you need:
Hour -> minute (map) -> Minute -> second (map) -> Second
Now connect the record to the right Calendar vertex. If the usual way to retrieve Log records is by hour you could link the Log records in the Hour. Example:
Year -> month (map) -> Month -> day (map) -> Day -> hour (map) -> Hour -> log (set) -> Log
The "log" property connects the Time Unit to the Log records. So to retrieve all the log of March 2012, 20th at 10am:
SELECT expand( month[3].day[20].hour[10].logs ) FROM Year WHERE year = "2012"
That could be used as starting point to retrieve only a sub-set of logs that satisfy certain rules. Example:
SELECT FROM (
SELECT expand( month[3].day[20].hour[10].logs ) FROM Year WHERE year = "2012"
) WHERE priority = 'critical'
That retrieves all the CRITICAL logs of March 2012, 20th at 10am.
Join multiple hours
If you need multiple hours/days/months as result set you can use the UNION function to create a unique result set:
SELECT expand( records ) from (
SELECT union( month[3].day[20].hour[10].logs, month[3].day[20].hour[11].logs ) AS records
FROM Year WHERE year = "2012"
)
In this example we create a union between the 10th and 11th hours. But what about extracting all the hours of a day without writing a huge query? The shortest way is using the Traverse. Below the Traverse to get all the hours of one day:
TRAVERSE hour FROM (
SELECT expand( month[3].day[20] ) FROM Year WHERE year = "2012"
)
So putting all together this query will extract all the logs of all the hours in a day:
SELECT expand( logs ) FROM (
SELECT union( logs ) AS logs FROM (
TRAVERSE hour FROM (
SELECT expand( month[3].day[20] ) FROM Year WHERE year = "2012"
)
)
)
Aggregate
Once you built up a Calendar in form of a Graph you can use it to store aggregated values and link them to the right Time Unit. Example: store all the winning ticket of Online Games. The record structure in our example is:
{
"date" : 12293289328932,
"win" : 10.34,
"machine" : "AKDJKD7673JJSH",
}
You can link this record to the closest Time Unit like in the example above, but you could sum all the records in the same Day and link it to the Day vertex. Example:
Create a new class to store the aggregated daily records:
CREATE CLASS DailyLog
Create the new record from an aggregation of the hour:
INSERT INTO DailyLog
SET win = (
SELECT SUM(win) AS win FROM Hour WHERE date BETWEEN '2012-03-20 10:00:00' AND '2012-03-20 11:00:00'
)
Link it in the Calendar graph assuming the previous command returned #23:45 as the RecordId of the brand new DailyLog record:
UPDATE (
SELECT expand( month[3].day[20] ) FROM Year WHERE year = "2012"
) ADD logs = #23:45
Chat Use Case
OrientDB allows modeling of rich and complex domains. If you want to develop a chat based application, you can use whatever you want to create the relationships between User and Room.
We suggest avoiding using Edges or Vertices connected with edges for messages. The best way is using the document API by creating one class per chat room, with no index, to have super fast access to last X messages. In facts, OrientDB stores new records in append only, and the @rid is auto generated as incrementing.
The 2 most common use cases in a chat are:
- writing a message in a chat room
- load last page of messages in a chat room
Create the initial schema
In order to work with the chat rooms, the rule of the thumb is creating a base abstract class ("ChatRoom") and then let to the concrete classes to represent individual ChatRooms.
Create the base ChatRoom class
create class ChatRoom
alter class ChatRoom abstract true
create property ChatRoom.date datetime
create property ChatRoom.text string
create property ChatRoom.user LINK OUser
Create a new ChatRoom
create class ItalianRestaurant extends ChatRoom
Class "ItalianRestaurant" will extend all the properties from ChatRoom.
Why creating a base class? Because you could always execute polymorphic queries that are cross-chatrooms, like get all the message from user "Luca":
select from ChatRoom where user.name = 'Luca'
Create a new message in the Chat Room
To create a new message in the chat room you can use this code:
public ODocument addMessage(String chatRoom, String message, OUser user) {
ODocument msg = new ODocument(chatRoom);
msg.field( "date", new Date() );
msg.field( "text", message );
msg.field( "user", user );
msg.save();
return msg;
}
Example:
addMessage("ItalianRestaurant", "Have you ever been at Ponza island?", database.getUser());
Retrieve last messages
You can easily fetch pages of messages ordered by date in descending order, by using the OrientDB's @rid
. Example:
select from ItalianRestaurant order by @rid desc skip 0 limit 50
You could write a generic method to access to a page of messages, like this:
public Iterable<ODocument> loadMessages(String chatRoom, fromLast, pageSize) {
return graph.getRawGraph().command("select from " + chatRoom + " order by @rid desc skip " + fromLast + " limit " + pageSize).execute();
}
Loading the 2nd (last) page from chat "ItalianRestaurant", would become this query (with pageSize = 50):
select from ItalianRestaurant order by @rid desc skip 50 limit 50
This is super fast and O(1) even with million of messages.
Limitations
Since OrientDB can handle only 32k clusters, you could have maximum 32k chat rooms. Unless you want to rewrite the entire FreeNode, 32k chat rooms will be more than enough for most of the cases.
However, if you need more than 32k chat rooms, the suggested solution is still using this approach, but with multiple databases (even on the same server, because one OrientDB Server instance can handle thousands of databases concurrently).
In this case you could use one database to handle all the metadata, like the following classes:
- ChatRoom, containing all the chatrooms, and the database where are stored. Example:
{ "@class": "ChatRoom", "description": "OrientDB public channel", "databaseName", "db1", "clusterName": "orientdb" }
- User, containing all the information about accounts with the edges to the ChatRoom vertices where they are subscribed
OrientDB cannot handle cross-database links, so when you want to know the message's author, you have to look up into the "Metadata" database by @RID (that is O(1)).
Key Value Use Case
OrientDB can also be used as a Key Value DBMS by using the super fast Indexes. You can have as many Indexes as you need.
HTTP
OrientDB RESTful HTTP protocol allows to talk with a OrientDB Server instance using the HTTP protocol and JSON. OrientDB supports also a highly optimized Binary protocol for superior performances.
Operations
To interact against OrientDB indexes use the four methods of the HTTP protocol in REST fashion:
- PUT, to create or modify an entry in the database
- GET, to retrieve an entry from the database. It's idempotent that means no changes to the database happen. Remember that in IE6 the URL can be maximum of 2,083 characters. Other browsers supports longer URLs, but if you want to stay compatible with all limit to 2,083 characters
- DELETE, to delete an entry from the database
Create an entry
To create a new entry in the database use the Index-PUT API.
Syntax: http://<server>:[<port>]/index/<index-name>/<key>
Example:
HTTP PUT: http://localhost:2480/index/customers/jay
{
"name" : "Jay",
"surname" : "Miner"
}
HTTP Response 204 is returned.
Retrieve an entry
To retrieve an entry from the database use the Index-GET API.
Syntax: http://<server>:[<port>]/index/<index-name>/<key>
Example:
HTTP GET: http://localhost:2480/index/customers/jay
HTTP Response 200 is returned with this JSON as payload:
{
"name" : "Jay",
"surname" : "Miner"
}
Remove an entry
To remove an entry from the database use the Index-DELETE API.
Syntax: http://<server>:[<port>]/index/<index-name>/<key>
Example:
HTTP DELETE: http://localhost:2480/index/customers/jay
HTTP Response 200 is returned
Step-by-Step tutorial
Before to start assure you've a OrientDB server up and running. In this example we'll use curl considering the connection to localhost to the default HTTP post 2480. The default "admin" user is used.
Create a new index
To use OrientDB as a Key/Value store we need a brand new manual index, let's call it "mainbucket". We're going to create it as UNIQUE because keys cannot be duplicated. If you can have multiple keys consider:
- creating the index as NOTUNIQUE
- leave it as UNIQUE but as value handle array of documents
Create the new manual unique index "mainbucket":
> curl --basic -u admin:admin localhost:2480/command/demo/sql -d "create index mainbucket UNIQUE STRING"
Response:
{ "result" : [
{ "@type" : "d" , "@version" : 0, "value" : 0, "@fieldTypes" : "value=l" }
]
}
Store the first entry
Below we're going to insert the first entry by using the HTTP PUT method passing "jay" as key in the URL and as value the entire document in form of JSON:
> curl --basic -u admin:admin -X PUT localhost:2480/index/demo/mainbucket/jay -d "{'@class': 'V', 'name':'Jay','surname':'Miner'}"
Response:
Key 'jay' correctly inserted into the index mainbucket.
Retrieve the entry just inserted
Below we're going to retrieve the entry we just entered by using the HTTP GET method passing "jay" as key in the URL:
> curl --basic -u admin:admin localhost:2480/index/demo/mainbucket/jay
Response:
[{
"@type" : "d" , "@rid" : "#3:477" , "@version" : 0,
"name" : "Jay",
"surname" : "Miner"
}]
Note that an array is always returned in case multiple records are associated to the same key (if NOTUNIQUE index is used). Look also at the document has been created with RID #3:477. You can load it directly if you know the RID. Remember to remove the # character. Example:
> curl --basic -u admin:admin localhost:2480/document/demo/3:477
Response:
{
"@type" : "d" , "@rid" : "#3:477" , "@version" : 0,
"name" : "Jay",
"surname" : "Miner"
}
Drop an index
Once finished drop the index "mainbucket" created for the example:
> curl --basic -u admin:admin localhost:2480/command/demo/sql -d "drop index mainbucket"
```json
Response:
```json
{ "result" : [
{ "@type" : "d" , "@version" : 0, "value" : 0, "@fieldTypes" : "value=l" }
]
}
Distributed queues use case
Implementing a persistent, distributed and transactional queue system using OrientDB is possible and easy. Besides the fact you don't need a specific API accomplish a queue, there are multiple approaches you can follow depending by your needs. The easiest way is using OrientDB SQL, so this works with any driver.
Create the queue class first:
create class queue
You could have one class per queue. Example of push operation:
insert into queue set text = "this is the first message", date = date()
Since OrientDB by default keeps the order of creation of records, a simple delete from the queue class with limit = 1 gives to you the perfect pop:
delete from queue return before limit 1
The "return before" allows you to have the deleted record content. If you need to peek the queue, you can just use the select:
select from queue limit 1
That's it. Your queue will be persistent, if you want transactional and running in cluster distributed.
#Import Tutorials
This section includes some import-to-OrientDB tutorials.
Tutorial: Importing the Open Beer Database into OrientDB
In this tutorial we will use the OrientDB's ETL module to import, as a graph, the Open Beer Database.
Note: You can access directly the converted database, result of this ETL tutorial, in the following ways:
-
Studio: in the login page press the "Cloud" button, put server's credential and press the download button from the "OpenBeer" row;
-
Direct Download: download the database from http://orientdb.com/public-databases/OpenBeer.zip and unzip it in a OpenBeer folder inside OrientDB's server "databases" directory.
The Open Beer Database
The Open Beer Database can be downloaded in CSV format from https://openbeerdb.com/. The following image shows its relational model:
Preliminary Steps
First, please create a new folder somewhere in your hard drive, and move into it. For this test we will assume /temp/openbeer
:
$ mkdir /temp/openbeer
$ cd /temp/openbeer
Download the Open Beer Database in CSV format
Download the Open Beer Database in CSV format and extract the archive:
$ curl http://openbeerdb.com/files/openbeerdb_csv.zip > openbeerdb_csv.zip
$ unzip openbeerdb_csv.zip
The archive consists of the following files:
beers.csv:
contains the beer recordsbreweries.csv:
contains the breweries recordsbreweries_geocode.csv
: contains the geocodes of the breweries. This file is not used in this Tutorialcategories.csv
: contains the beer categoriesstyles.csv
: contains the beer styles
Install OrientDB
Download and install OrientDB:
$ wget https://repo1.maven.org/maven2/com/orientechnologies/orientdb-community/3.2.33/orientdb-community-3.2.33.tar.gz -O orientdb-community-3.2.33.tar.gz
$ tar xvf orientdb-community-3.2.33.tar.gz
For more information on how to install OrientDB, please refer to the Installation section.
Graph Data Model
Before starting the ETL process it's important to understand how the Open Beer Database can be modeled as a graph.
The relational model of the Open Beer Database can be easily converted to a graph model, as shown below:
The model above consists of the following nodes (or vertices) and relationships (or edges):
- Nodes: Beer, Category, Style, Brewery;
- Relationships: HasCategory, HasStyle, HasBrewery.
For more informations on the Graph Model in OrientDB, please refer to the Graph Model section.
ETL Process
The ETL module for OrientDB provides support for moving data to and from OrientDB databases using Extract, Transform and Load processes.
The ETL module consists of a script, oetl.sh
, that takes in input a single JSON configuration file.
For more information on the ETL module, please refer to the ETL section.
IMPORTANT: If you use ETL in PLOCAL, make sure you do not have a running server on the same folder, you will an error otherwise:
Database is locked by another process, please shutdown process and try again
Import Beer Categories
The following are the first two lines of the categories.csv
file:
"id","cat_name","last_mod"
"1","British Ale","2010-10-24 13:50:10"
In order to import this file in OrientDB, we have to create the following file as categories.json
:
{
"source": { "file": { "path": "/temp/openbeer/openbeerdb_csv/categories.csv" } },
"extractor": { "csv": {} },
"transformers": [
{ "vertex": { "class": "Category" } }
],
"loader": {
"orientdb": {
"dbURL": "plocal:../databases/openbeerdb",
"dbType": "graph",
"classes": [
{"name": "Category", "extends": "V"}
], "indexes": [
{"class":"Category", "fields":["id:integer"], "type":"UNIQUE" }
]
}
}
}
To import it into OrientDB, please move into the "bin" directory of the OrientDB distribution:
$ cd orientdb-community-2.2.8/bin
and run OrientDB ETL:
$ ./oetl.sh /temp/openbeer/categories.json
OrientDB etl v.2.0.9 (build @BUILD@) www.orientechnologies.com
BEGIN ETL PROCESSOR
END ETL PROCESSOR
+ extracted 12 rows (0 rows/sec) - 12 rows -> loaded 11 vertices (0 vertices/sec) Total time: 77ms [0 warnings, 0 errors]
Import Beer Styles
Now let's import the Beer Styles. These are the first two lines of the styles.csv
file:
"id","cat_id","style_name","last_mod"
"1","1","Classic English-Style Pale Ale","2010-10-24 13:53:31"
In this case we will correlate the Style with the Category created earlier.
This is the styles.json
to use with OrientDB ETL for the next step:
{
"source": { "file": { "path": "/temp/openbeer/openbeerdb_csv/styles.csv" } },
"extractor": { "csv": {} },
"transformers": [
{ "vertex": { "class": "Style" } },
{ "edge": { "class": "HasCategory", "joinFieldName": "cat_id", "lookup": "Category.id" } }
],
"loader": {
"orientdb": {
"dbURL": "plocal:../databases/openbeerdb",
"dbType": "graph",
"classes": [
{"name": "Style", "extends": "V"},
{"name": "HasCategory", "extends": "E"}
], "indexes": [
{"class":"Style", "fields":["id:integer"], "type":"UNIQUE" }
]
}
}
}
Now, to import the styles, please execute the following command:
$ ./oetl.sh /temp/openbeer/styles.json
OrientDB etl v.2.0.9 (build @BUILD@) www.orientechnologies.com
BEGIN ETL PROCESSOR
END ETL PROCESSOR
+ extracted 142 rows (0 rows/sec) - 142 rows -> loaded 141 vertices (0 vertices/sec) Total time: 498ms [0 warnings, 0 errors]
Import Breweries
Now it's time for the Breweries. These are the first two lines of the breweries.csv
file:
"id","name","address1","address2","city","state","code","country","phone","website","filepath","descript","last_mod"
"1","(512) Brewing Company","407 Radam, F200",,"Austin","Texas","78745","United States","512.707.2337","http://512brewing.com/",,"(512) Brewing Company is a microbrewery located in the heart of Austin that brews for the community using as many local, domestic and organic ingredients as possible.","2010-07-22 20:00:20"
Breweries have no outgoing relations with other entities, so this is a plain import similar to the one we did for the categories.
This is the breweries.json
to use with OrientDB ETL for the next step:
{
"source": { "file": { "path": "/temp/openbeer/openbeerdb_csv/breweries.csv" } },
"extractor": { "csv": {} },
"transformers": [
{ "vertex": { "class": "Brewery" } }
],
"loader": {
"orientdb": {
"dbURL": "plocal:../databases/openbeerdb",
"dbType": "graph",
"classes": [
{"name": "Brewery", "extends": "V"}
], "indexes": [
{"class":"Brewery", "fields":["id:integer"], "type":"UNIQUE" }
]
}
}
}
Run the import for breweries:
$ ./oetl.sh /temp/openbeer/breweries.json
OrientDB etl v.2.0.9 (build @BUILD@) www.orientechnologies.com
BEGIN ETL PROCESSOR
END ETL PROCESSOR
+ extracted 1.395 rows (0 rows/sec) - 1.395 rows -> loaded 1.394 vertices (0 vertices/sec) Total time: 830ms [0 warnings, 0 errors]
Import Beers
Now it's time for the last and most important file: the Beers! These are the first two lines of the beers.csv
file:
"id","brewery_id","name","cat_id","style_id","abv","ibu","srm","upc","filepath","descript","last_mod",,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,,
"1","812","Hocus Pocus","11","116","4.5","0","0","0",,"Our take on a classic summer ale. A toast to weeds, rays, and summer haze. A light, crisp ale for mowing lawns, hitting lazy fly balls, and communing with nature, Hocus Pocus is offered up as a summer sacrifice to clodless days.
As you can see each beer is connected to other entities through the following fields:
brewery_id
-> Brewerycat_id
-> Categorystyle_id
-> Style
This is the beers.json
to use with OrientDB ETL for the next step:
{
"config" : { "haltOnError": false },
"source": { "file": { "path": "/temp/openbeer/openbeerdb_csv/beers.csv" } },
"extractor": { "csv": { "columns": ["id","brewery_id","name","cat_id","style_id","abv","ibu","srm","upc","filepath","descript","last_mod"],
"columnsOnFirstLine": true } },
"transformers": [
{ "vertex": { "class": "Beer" } },
{ "edge": { "class": "HasCategory", "joinFieldName": "cat_id", "lookup": "Category.id" } },
{ "edge": { "class": "HasBrewery", "joinFieldName": "brewery_id", "lookup": "Brewery.id" } },
{ "edge": { "class": "HasStyle", "joinFieldName": "style_id", "lookup": "Style.id" } }
],
"loader": {
"orientdb": {
"dbURL": "plocal:../databases/openbeerdb",
"dbType": "graph",
"classes": [
{"name": "Beer", "extends": "V"},
{"name": "HasCategory", "extends": "E"},
{"name": "HasStyle", "extends": "E"},
{"name": "HasBrewery", "extends": "E"}
], "indexes": [
{"class":"Beer", "fields":["id:integer"], "type":"UNIQUE" }
]
}
}
}
Run the final import for beers:
$ ./oetl.sh /temp/openbeer/beers.json
OrientDB etl v.2.0.9 (build @BUILD@) www.orientechnologies.com
BEGIN ETL PROCESSOR
...
+ extracted 5.862 rows (1.041 rows/sec) - 5.862 rows -> loaded 4.332 vertices (929 vertices/sec) Total time: 10801ms [0 warnings, 27 errors]
END ETL PROCESSOR
Note: the 27 errors are due to the 27 wrong content lines that have no id.
Some Queries and Visualizations
Now that the database has been imported we can execute some queries and create some visualizations.
The following are some ways we can use to access the newly imported OpenBeer
database:
- Console
- Gremlin Console
- Studio
- APIs & Drivers
- some external tools, like Gephi
- some external visualization libraries for graph rendering
If we want to query all Category vertices we can execute the following query:
SELECT * FROM Category
The following is the visualization we can create using the Studio's Graph Editor:
If we want to find all nodes directly connected to a specific beer (e.g. the beer Petrus Dubbel Bruin Ale) with either an incoming or outgoing relationship, we can use a query like the following:
SELECT EXPAND( BOTH() ) FROM Beer WHERE name = 'Petrus Dubbel Bruin Ale'
Alternatively, we can use the MATCH syntax:
MATCH {class: Beer, where: (name = 'Petrus Dubbel Bruin Ale')}--{as: n} RETURN $pathelements
If we execute the first query in the Browse tab of Studio we get the following result, from where we can see that there are three nodes connected to this beer, having @rid 11:4, 14:262 and 12:59:
We can send the result of this SELECT
query to the Graph Editor by clicking the icon "Send to Graph", or create a new visualization directly from the Graph Editor.
The following is the visualization of the MATCH
query above, executed directly on the Graph Editor:
The same resultset can be visualized using an external graph library. For instance, the following graph has been obtained using the library vis.js where the input visjs dataset has been created with a java program created using the OrientDB's Java Graph API:
We can also query bigger portions of the graph. For example, to query all beer Category nodes and for each of them all the connected Style nodes, we can use a MATCH query like the following:
MATCH
{class: Category, as: category}-HasCategory-{class: Style, as: style}
RETURN $elements
The following is the visualization of the MATCH
query above in the Graph Editor:
while the following is a visualization created for the same recordset using the library vis.js:
Tutorial: Importing the northwind Database from Neo4j
In this tutorial we will use the Neo4j to OrientDB Importer to import the Neo4j northwind example database into OrientDB.
For general information on the possible Neo4j to OrientDB migration strategies, please refer to the Import from Neo4j section.
Neo4j and Cypher are registered trademark of Neo Technology, Inc.
Preparing for the migration
Please download and install OrientDB:
$ wget https://repo1.maven.org/maven2/com/orientechnologies/orientdb-community-tp2/3.2.33/orientdb-community-tp2-3.2.33.tar.gz -O orientdb-community-tp2-3.2.33.tar.gz
$ tar xvf orientdb-community-tp2-3.2.33.tar.gz
Download and install the Neo4j to OrientDB Importer:
$ wget http://central.maven.org/maven2/com/orientechnologies/orientdb-neo4j-importer/3.2.33/orientdb-neo4j-importer-3.2.33.tar.gz
$ tar xfv orientdb-neo4j-importer-3.2.33.tar.gz -C orientdb-community-3.2.33 --strip-components=1
For further information on the OrientDB's installation, please refer to this section.
For further information on the Neo4j to OrientDB Importer installation, please refer to this section.
Starting the migration
Assuming that:
-
/home/santo/neo4j/neo4j-community-3.0.7/lib
is the full path to the directory that includes the Neo4j's libraries -
/home/santo/data/graph.db_northwind
is the full path to the directory that contains the Neo4j's northwind database -
/home/santo/orientdb/orientdb-community-2.2.12/databases/northwind_import
is the full path to the directory where you would like to migrate the northwind database -
that no Neo4j and OrientDB servers are running on those directories
you can import the northwind database with a command similar to the following:
./orientdb-neo4j-importer.sh \
-neo4jlibdir /home/santo/neo4j/neo4j-community-3.0.7/lib \
-neo4jdbdir /home/santo/neo4j/data/graph.db_northwind \
-odbdir /home/santo/orientdb/orientdb-community-2.2.12/databases/northwind_import
For further information on how to use the Neo4j to OrientDB Importer, please refer to this section.
Migration output
The following is the output that is written by the Neo4j to OrientDB Importer during the northwind
database migration:
Neo4j to OrientDB Importer v.2.2.12-SNAPSHOT - Copyrights (c) 2016 OrientDB LTD
WARNING: 'o' option not found. Defaulting to 'false'.
Please make sure that there are no running servers on:
'/home/santo/neo4j/data/graph.db_northwind' (Neo4j)
and:
'/home/santo/orientdb/orientdb-community-2.2.12/databases/northwind_import' (OrientDB)
Initializing Neo4j...Done
Initializing OrientDB...Done
Importing Neo4j database:
'/home/santo/neo4j/data/graph.db_northwind'
into OrientDB database:
'/home/santo/orientdb/orientdb-community-2.2.12/databases/northwind_import'
Getting all Nodes from Neo4j and creating corresponding Vertices in OrientDB...
1035 OrientDB Vertices have been created (100% done)
Done
Creating internal Indices on property 'Neo4jNodeID' on all OrientDB Vertices Classes...
5 OrientDB Indices have been created (100% done)
Done
Getting all Relationships from Neo4j and creating corresponding Edges in OrientDB...
3139 OrientDB Edges have been created (100% done)
Done
Getting Constraints from Neo4j and creating corresponding ones in OrientDB...
0 OrientDB Indices have been created
Done
Getting Indices from Neo4j and creating corresponding ones in OrientDB...
5 OrientDB Indices have been created (100% done)
Done
Import completed!
Shutting down OrientDB...Done
Shutting down Neo4j...Done
===============
Import Summary:
===============
- Found Neo4j Nodes : 1035
-- With at least one Label : 1035
--- With multiple Labels : 0
-- Without Labels : 0
- Imported OrientDB Vertices : 1035 (100%)
- Found Neo4j Relationships : 3139
- Imported OrientDB Edges : 3139 (100%)
- Found Neo4j Constraints : 0
- Imported OrientDB Constraints (Indices created) : 0
- Found Neo4j (non-constraint) Indices : 5
- Imported OrientDB Indices : 5 (100%)
- Additional created Indices (on vertex properties 'Neo4jNodeID') : 5
- Total Import time: : 29 seconds
-- Initialization time : 7 seconds
-- Time to Import Nodes : 6 seconds (181.67 nodes/sec)
-- Time to Import Relationships : 7 seconds (459.79 rels/sec)
-- Time to Import Constraints and Indices : 4 seconds (1.21 indices/sec)
-- Time to create internal Indices (on vertex properties 'Neo4jNodeID') : 4 seconds (1.22 indices/sec)
Connecting to the newly imported Database
General information on how to connect to a newly imported database can be found in this section.
The following is a partial visualization of the northwind database done with the Graph Editor included in the OrientDB's Studio tool:
As you can see from the Limit field, the visualization is limited to 200 vertices.
The following, instead, is the graph returned by the following MATCH
query (the query returns all nodes connected to the Order with orderID
10344):
MATCH {class: Order, where: (orderID = 10344)}--{as: n} RETURN $pathelements
From Studio's Schema Manager, you can check all imported Vertex Classes (node Labels in Neo4j), Edge Classes (Relationship Types in Neo4j), and Indexes:
V
and E
are special classes: they include all Vertices and all Edges.
This is a legacy strategy to migrate from Neo4j. The new strategy is to migrate using the Neo4j to OrientDB Importer. |
Tutorial: Importing the movie Database from Neo4j
In this tutorial we will follow the steps described in the Import from Neo4j using GraphML section to import the Neo4j's movie example database into OrientDB.
We will also provide some examples of queries using the OrientDB's MATCH syntax, making a comparison with the corresponding Neo4j's Cypher query language.
For general information on the possible Neo4j to OrientDB migration strategies, please refer to the Import from Neo4j section.
Neo4j and Cypher are registered trademark of Neo Technology, Inc.
Exporting from Neo4j
Assuming you have already downloaded and unpacked the Neo4j Shell Tools, and restarted the Neo4j Server, as described in the Section Exporting GraphML, you can export the movie database using neo4j-shell
with a command like the following one:
D:\neo4j\neo4j-community-3.0.6\bin>neo4j-shell.bat
Welcome to the Neo4j Shell! Enter 'help' for a list of commands
NOTE: Remote Neo4j graph database service 'shell' at port 1337
neo4j-sh (?)$ export-graphml -t -o d:/movie.graphml
Wrote to GraphML-file d:/movies.graphml 0. 100%: nodes = 171 rels = 253 properties = 564 time 270 ms total 270 ms
In the example above the exported movie graph is stored under D:\movie.graphml
.
Importing into OrientDB
In this tutorial we will import in OrientDB the file movie.graphml
using the OrientDB's Console. For other GraphML import methods, please refer to the section Importing GraphML.
The OrientDB's Console output generated during the import process is similar to the following (note that first we create a movie database using the command CREATE DATABASE
, and then we do the actual import using the command IMPORT DATABASE
):
D:\orientdb\orientdb-enterprise-2.2.8\bin>console.bat
OrientDB console v.2.2.8-SNAPSHOT (build 2.2.x@r39259e190e16045fe1425b1c0485f8562fca055b; 2016-08-23 14:38:49+0000) www.orientdb.com
Type 'help' to display all the supported commands.
Installing extensions for GREMLIN language v.2.6.0
orientdb> CREATE DATABASE PLOCAL:D:/orientdb/orientdb-enterprise-2.2.8/databases/movie
Creating database [PLOCAL:D:/orientdb/orientdb-enterprise-2.2.8/databases/movie] using the storage type [PLOCAL]...
Database created successfully.
Current database is: PLOCAL:D:/orientdb/orientdb-enterprise-2.2.8/databases/movie
orientdb {db=movie}> IMPORT DATABASE D:/movie.graphml
Importing GRAPHML database from D:/movie.graphml with options ()...
Done: imported 171 vertices and 253 edges
orientdb {db=movie}>
As you can see from the output above, as a result of the import 171 vertices and 253 edges have been created in OrientDB. This is exactly the same number of nodes and relationships exported from Neo4j.
For more tips and tricks related to the import process, please refer to this section.
Query Comparison
Once the movie database has been imported into OrientDB, you may use several ways to access its data.
The MATCH
syntax and the tool Studio can be used, for instance, in a similar way to the Neo4j's Cypher and Browser.
The following sections include a comparison of the Neo4j's Cypher and OrientDB's MATCH
syntax for some queries that you can execute against the movie database.
Find the actor named "Tom Hanks"
Neo4j's Cypher:
MATCH (tom:Person {name: "Tom Hanks"})
RETURN tom
OrientDB's MATCH:
MATCH {class: Person, as: tom, where: (name = 'Tom Hanks')}
RETURN $pathElements
Find the movie with title "Cloud Atlas"
Neo4j's Cypher:
MATCH (cloudAtlas:Movie {title: "Cloud Atlas"})
RETURN cloudAtlas
OrientDB's MATCH:
MATCH {class: Movie, as: cloudAtlas, where: (title = 'Cloud Atlas')}
RETURN $pathElements
Find 10 people
Neo4j's Cypher:
MATCH (people:Person)
RETURN people.name
LIMIT 10
OrientDB's MATCH:
MATCH {class: Person, as: people}
RETURN people.name
LIMIT 10
Find the movies released in the 1990s
Neo4j's Cypher:
MATCH (nineties:Movie)
WHERE nineties.released > 1990 AND nineties.released < 2000
RETURN nineties.title
OrientDB's MATCH:
MATCH {class: Movie, as: nineties, WHERE: (released > 1990 AND released < 2000 )}
RETURN nineties.title
List all Tom Hanks movies
Neo4j's Cypher:
MATCH (tom:Person {name: "Tom Hanks"})-[:ACTED_IN]->(tomHanksMovies)
RETURN tom, tomHanksMovies
OrientDB's MATCH:
MATCH {class: Person, as: tom, where: (name = 'Tom Hanks')}-ACTED_IN->{as: tomHanksMovies}
RETURN $pathElements
Find out who directed "Cloud Atlas"
Neo4j's Cypher:
MATCH (cloudAtlas {title: "Cloud Atlas"})<-[:DIRECTED]-(directors)
RETURN directors.name
OrientDB's MATCH:
MATCH {class: Movie, as: cloudAtlas, where: (title = 'Cloud Atlas')}<-DIRECTED-{as: directors}
RETURN directors.name
Find Tom Hanks' co-actors
Neo4j's Cypher:
MATCH (tom:Person {name:"Tom Hanks"})-[:ACTED_IN]->(m)<-[:ACTED_IN]-(coActors)
RETURN DISTINCT coActors.name
OrientDB's MATCH:
MATCH {class: Person, as: tom, where: (name = 'Tom Hanks')}-ACTED_IN->{as: m}<-ACTED_IN-{class: Person,as: coActors}
RETURN coActors.name
Find how people are related to "Cloud Atlas"
Neo4j's Cypher:
MATCH (people:Person)-[relatedTo]-(:Movie {title: "Cloud Atlas"})
RETURN people.name, Type(relatedTo), relatedTo
OrientDB's MATCH:
MATCH {class: Person, as: people}--{as: m, where: (title = 'Cloud Atlas')}
RETURN $pathElements
General Information
This Chapter includes some General Information on OrientDB.
Supported Data Types
OrientDB supports several data types natively. Below is the complete table.
#id | Type | SQL type | Description | Java type | Minimum Maximum | Auto-conversion from/to |
---|---|---|---|---|---|---|
0 | Boolean | BOOLEAN | Handles only the values True or False | java.lang.Boolean or boolean | 0 1 | String |
1 | Integer | INTEGER | 32-bit signed Integers | java.lang.Integer or int | -2,147,483,648 +2,147,483,647 | Any Number, String |
2 | Short | SHORT | Small 16-bit signed integers | java.lang.Short or short | -32,768 32,767 | Any Number, String |
3 | Long | LONG | Big 64-bit signed integers | java.lang.Long or long | -263 +263-1 | Any Number, String |
4 | Float | FLOAT | Decimal numbers | java.lang.Float or float | 2-149 (2-2-23)*2127 | Any Number, String |
5 | Double | DOUBLE | Decimal numbers with high precision | java.lang.Double or double | 2-1074 (2-2-52)*21023 | Any Number, String |
6 | Datetime | DATETIME | Any date with the precision up to milliseconds. To know more about it, look at Managing Dates | java.util.Date | - 1002020303 | Date, Long, String |
7 | String | STRING | Any string as alphanumeric sequence of chars | java.lang.String | - - | - |
8 | Binary | BINARY | Can contain any value as byte array | byte[] | 0 2,147,483,647 | String |
9 | Embedded | EMBEDDED | The Record is contained inside the owner. The contained Record has no Record ID | ORecord | - - | ORecord |
10 | Embedded list | EMBEDDEDLIST | The Records are contained inside the owner. The contained records have no Record ID's and are reachable only by navigating the owner record | List<Object> | 0 41,000,000 items | String |
11 | Embedded set | EMBEDDEDSET | The Records are contained inside the owner. The contained Records have no Record ID and are reachable only by navigating the owner record | Set<Object> | 0 41,000,000 items | String |
12 | Embedded map | EMBEDDEDMAP | The Records are contained inside the owner as values of the entries, while the keys can only be Strings. The contained ords e no Record IDs and are reachable only by navigating the owner Record | Map<String, ORecord> | 0 41,000,000 items | Collection<? extends ORecord<?>> , String |
13 | Link | LINK | Link to another Record. It's a common one-to-one relationship | ORID , <? extends ORecord> | 1:-1 32767:2^63-1 | String |
14 | Link list | LINKLIST | Links to other Records. It's a common one-to-many relationship where only the Record IDs are stored | List<? extends ORecord | 0 41,000,000 items | String |
15 | Link set | LINKSET | Links to other Records. It's a common one-to-many relationship | Set<? extends ORecord> | 0 41,000,000 items | Collection<? extends ORecord> , String |
16 | Link map | LINKMAP | Links to other Records as value of the entries, while keys can only be Strings. It's a common One-to-Many Relationship. Only the Record IDs are stored | Map<String, | 0 41,000,000 items | String |
17 | Byte | BYTE | Single byte. Useful to store small 8-bit signed integers | java.lang.Byte or byte | -128 +127 | Any Number, String |
18 | Transient | TRANSIENT | Any value not stored on database | |||
19 | Date | DATE | Any date as year, month and day. To know more about it, look at Managing Dates | java.util.Date | - | Date, Long, String |
20 | Custom | CUSTOM | used to store a custom type providing the marshall and unmarshall methods | OSerializableStream | 0 X | - |
21 | Decimal | DECIMAL | Decimal numbers without rounding | java.math.BigDecimal | ? ? | Any Number, String |
22 | LinkBag | LINKBAG | List of Record IDs as spec RidBag | ORidBag | ? ? | - |
23 | Any | ANY | Not determinated type, used to specify Collections of mixed type, and null | - | - | - |
Schema
While OrientDb can work in a schema-less mode, you may find it necessary at times to enforce a schema on your data model. OrientDB supports both schema-full and schema-hybrid solutions.
In the case of schema-hybrid mode, you only set constraints for certain fields and leave the user to add custom fields to the record. This mode occurs at a class level, meaning that you can have an Employee
class as schema-full and an EmployeeInformation
class as schema-less.
- Schema-full Enables strict-mode at a class-level and sets all fields as mandatory.
- Schema-less Enables classes with no properties. Default is non-strict-mode, meaning that records can have arbitrary fields.
- Schema-hybrid Enables classes with some fields, but allows records to define custom fields. This is also sometimes called schema-mixed.
NOTE Changes to the schema are not transactional. You must execute these commands outside of a transaction.
You can access the schema through SQL or through the Java API. Examples here use the latter. To access the schema API in Java, you need the Schema instance of the database you want to use. For example,
OSchema schema = database.getMetadata().getSchema();
Class
OrientDB draws from the Object Oriented programming paradigm in the concept of the Class. A class is a type of record. In comparison to Relational database systems, it is most similar in conception to the table.
Classes can be schema-less, schema-full or schema-hybrid. They can inherit from other classes, shaping a tree of classes. In other words, a sub-class extends the parent class, inheriting all attributes.
Each class has its own clusters. By default, these clusters are logical, but they can also be physical. A given class must have at least one cluster defined as its default, but it can support multiple clusters. OrientDB writes new records into the default cluster, but always reads from all defined clusters.
When you create a new class, OrientDB creates a default physical cluster that uses the same name as the class, but in lowercase.
Creating Persistent Classes
Classes contain one or more properties. This mode is similar to the classical model of the Relational database, where you must define tables before you can begin to store records.
To create a persistent class in Java, use the createClass()
method:
OClass account = database.getMetadata().getSchema().createClass("Account");
This method creates the class Account
on the database. It simultaneously creates the physical cluster account
, to provide storage for records in the class Account
.
Getting Persistent Classes
With the new persistent class created, you may also need to get its contents.
To retrieve a persistent class in Java, use the getClass()
method:
OClass account = database.getMetadata().getSchema().getClass("Account");
This method retrieves from the database the persistent class Account
. If the query finds that the Account
class does not exist, it returns NULL
.
Dropping Persistent Classes
In the event that you no longer want the class, you can drop, or delete, it from the database.
To drop a persistent class in Java, use the OSchema.dropClass()
method:
database.getMetadata().getSchema().dropClass("Account");
This method drops the class Account
from your database. It does not delete records that belong to this class unless you explicitly ask it to do so:
database.command(new OCommandSQL("DELETE FROM Account")).execute();
database.getMetadata().getSchema().dropClass("Account");
Constraints
Working in schema-full mode requires that you set the strict mode at the class-level, by defining the setStrictMode()
method to TRUE
. In this case, records of that class cannot have undefined properties.
Properties
In OrientDB, a property is a field assigned to a class. For the purposes of this tutorial, consider Property and Field as synonymous.
Creating Class Properties
After you create a class, you can define fields for that class. To define a field, use the createProperty()
method.
OClass account = database.getMetadata().getSchema().createClass("Account");
account.createProperty("id", OType.Integer);
account.createProperty("birthDate", OType.Date);
These lines create a class Account
, then defines two properties id
and birthDate
. Bear in mind that each field must belong to one of the supported types. Here these are the integer and date types.
Dropping Class Properties
In the event that you would like to remove properties from a class you can do so using the dropProperty()
method under OClass
.
database.getMetadata().getSchema().getClass("Account").dropProperty("name");
When you drop a property from a class, it does not remove records from that class unless you explicitly ask for it, using the UPDATE... REMOVE
statements. For instance,
database.getMetadata().getSchema().getClass("Account").dropProperty("name");
database.command(new OCommandSQL("UPDATE Account REMOVE name")).execute();
The first method drops the property from the class. The second updates the database to remove the property.
Relationships
OrientDB supports two types of relationships: referenced and embedded.
Referenced Relationships
In the case of referenced relationships, OrientDB uses a direct link to the referenced record or records. This allows the database to avoid the costly JOIN
operations used by Relational databases.
customer
Record A -------------> Record B
CLASS=Invoice CLASS=Customer
RID=5:23 RID=10:2
In the example, Record A contains the reference to Record B in the property customer
. Both records are accessible by any other records since each has a Record ID.
1:1 and n:1 Reference Relationships
In one to one and many to one relationships, the reference relationship is expressed using the LINK
type. For instance.
OClass customer= database.getMetadata().getSchema().createClass("Customer");
customer.createProperty("name", OType.STRING);
OClass invoice = database.getMetadata().getSchema().createClass("Invoice");
invoice.createProperty("id", OType.INTEGER);
invoice.createProperty("date", OType.DATE);
invoice.createProperty("customer", OType.LINK, customer);
Here, records of the class Invoice
link to a record of the class Customer
, through the field customer
.
1:n and n:n Reference Relationships.
In one to many and many to many relationships, OrientDB expresses the referenced relationship using collections of links.
LINKLIST
An ordered list of links.LINKSET
An unordered set of links, that does not accept duplicates.LINKMAP
An ordered map of links, with a string key. It does not accept duplicate keys.
For example,
OClass orderItem = db.getMetadata().getSchema().createClass("OrderItem");
orderItem.createProperty("id", OType.INTEGER);
orderItem.createProperty("animal", OType.LINK, animal);
OClass order = db.getMetadata().getSchema().createClass("Order");
order.createProperty("id", OType.INTEGER);
order.createProperty("date", OType.DATE);
order.createProperty("items", OType.LINKLIST, orderItem);
Here, you have two classes: Order
and OrderItem
and a 1:n referenced relationship is created between them.
Embedded Relationships
In the case of embedded relationships, OrientDB contains the relationship within the record. Embedded relationships are stronger than referenced relationships, but the embedded record does not have its own Record ID. Because of this, you cannot reference them directly through other records. The relationship is only accessible through the container record. If the container record is deleted, then the embedded record is also deleted.
address
Record A <>----------> Record B
CLASS=Account CLASS=Address
RID=5:23 NO RID!
Here, Record A contains the entirety of Record B in the property address
. You can only reach Record B by traversing the container, Record A.
orientdb> SELECT FROM Account WHERE Address.city = 'Rome'
1:1 and n:1 Embedded Relationships
For one to one and many to one embedded relationships, OrientDB uses links of the EMBEDDED
type. For example,
OClass address = database.getMetadata().getSchema().createClass("Address");
OClass account = database.getMetadata().getSchema().createClass("Account");
account.createProperty("id", OType.INTEGER);
account.createProperty("birthDate", OType.DATE);
account.createProperty("address", OType.EMBEDDED, address);
Here, records of the class Account
embed records for the class Address
.
1:n and n:n Embedded Relationships
In the case of one to many and many to many relationships, OrientDB sues a collection embedded link types:
EMBEDDEDLIST
An ordered list of records.EMBEDDEDSET
An unordered set of records. It doesn't accept duplicates.EMBEDDEDMAP
An ordered map of records as key-value pairs. It doesn't accept duplicate keys.
For example,
OClass orderItem = db.getMetadata().getSchema().createClass("OrderItem");
orderItem.createProperty("id", OType.INTEGER);
orderItem.createProperty("animal", OType.LINK, animal);
OClass order = db.getMetadata().getSchema().createClass("Order");
order.createProperty("id", OType.INTEGER);
order.createProperty("date", OType.DATE);
order.createProperty("items", OType.EMBEDDEDLIST, orderItem);
This establishes a one to many relationship between the classes Order
and OrderItem
.
Constraints
OrientDB supports a number of constraints for each field. For more information on setting constraints, see the ALTER PROPERTY
command.
- Minimum Value:
setMin()
The field accepts a string, because it works also for date ranges. - Maximum Value:
setMax()
The field accepts a string, because it works also for date rangers. - Mandatory:
setMandatory()
This field is required. - Read Only:
setReadonly()
This field cannot update after being created. - Not Null:
setNotNull()
This field cannot be null. - Unique: This field doesn't allow duplicates or speedup searches.
- Regex: This field must satisfy Regular Expressions
For example,
profile.createProperty("nick", OType.STRING).setMin("3").setMax("30").setMandatory(true).setNotNull(true);
profile.createIndex("nickIdx", OClass.INDEX_TYPE.UNIQUE, "nick"); // Creates unique constraint
profile.createProperty("name", OType.STRING).setMin("3").setMax("30");
profile.createProperty("surname", OType.STRING).setMin("3").setMax("30");
profile.createProperty("registeredOn", OType.DATE).setMin("2010-01-01 00:00:00");
profile.createProperty("lastAccessOn", OType.DATE).setMin("2010-01-01 00:00:00");
Indices as Constraints
To define a property value as unique, use the UNIQUE
index constraint. For example,
profile.createIndex("EmployeeId", OClass.INDEX_TYPE.UNIQUE, "id");
You can also constrain a group of properties as unique by creating a composite index made from multiple fields. For instance,
profile.createIndex("compositeIdx", OClass.INDEX_TYPE.NOTUNIQUE, "name", "surname");
For more information about indexes look at Index guide.
Inheritance
Unlike many Object-relational mapping tools, OrientDB does not split documents between different classes. Each document resides in one or a number of clusters associated with its specific class. When you execute a query against a class that has subclasses, OrientDB searches the clusters of the target class and all subclasses.
Declaring Inheritance in Schema
In developing your application, bear in mind that OrientDB needs to know the class inheritance relationship. This is an abstract concept that applies to both POJO's and Documents.
For example,
OClass account = database.getMetadata().getSchema().createClass("Account");
OClass company = database.getMetadata().getSchema().createClass("Company").setSuperClass(account);
Using Polymorphic Queries
By default, OrientDB treats all queries as polymorphic. Using the example above, you can run the following query from the console:
orientdb> SELECT FROM Account WHERE name.toUpperCase() = 'GOOGLE'
This query returns all instances of the classes Account
and Company
that have a property name that matches Google
.
How Inheritance Works
Consider an example, where you have three classes, listed here with the cluster identifier in the parentheses.
Account(10) <|--- Company (13) <|--- OrientTechnologiesGroup (27)
By default, OrientDB creates a separate cluster for each class. It indicates this cluster by the defaultClusterId
property in the class OClass
and indicates the cluster used by default when not specified. However, the class OClass
has a property clusterIds
, (as int[]
), that contains all the clusters able to contain the records of that class. clusterIds
and defaultClusterId
are the same by default.
When you execute a query against a class, OrientDB limits the result-sets to only the records of the clusters contained in the clusterIds
property. For example,
orientdb> SELECT FROM Account WHERE name.toUpperCase() = 'GOOGLE'
This query returns all the records with the name property set to GOOGLE
from all three classes, given that the base class Account
was specified. For the class Account
, OrientDB searches inside the clusters 10
, 13
and 27
, following the inheritance specified in the schema.
Concurrency
OrientDB uses an optimistic approach to concurrency. Optimistic Concurrency Control, or OCC assumes that multiple transactions can compete frequently without interfering with each other. It's very important that you don't share instances of databases, graphs, records, documents, vertices and edges between threads because they are non thread-safe. For more information look at Multi-Threading.
How does it work?
Consider the following scenario, where 2 clients, A and B, want to update the amount of a bank account:
Client A Client B
| |
(t1) |
Read record #13:22 |
amount is 100 (t2)
| Read record #13:22
(t3) amount is 100
Update record #13:22 |
set amount = amount + 10 (t4)
| Update record #13:22
| set amount = amount + 10
| |
Client A (t1) and B (t2) read the record #13:22 and both receive the last amount as USD 100. Client A updates the amount by adding USD 10 (t3), then the Client B is trying to do the same thing: updates the amount by adding USD 10. Here is the problem: Client B is doing an operation based on current information: the amount was USD 100. But at the moment of update, such information is changed (by Client A on t3), so the amount is USD 110 in the database. Should the update succeed by setting the new amount to USD 120?
In some cases this could be totally fine, in others not. It depends by the use case. For example, in your application there could be a logic where you are donating USD 10 to all the accounts where the amount is <=100. The owner of the account behind the record #13:22 is more lucky than the others, because it receives the donation even if it has USD 110 at that moment.
For this reason in OrientDB when this situation happens a OConcurrentModificationException
exception is thrown, so the application can manage it properly. Usually the 3 most common strategies to handle this exceptions are:
- Retry doing the same operation by reloading the record #13:22 first with the updated amount
- Ignore the change, because the basic condition is changed
- Propagate the exception to the user, so he can decide what to do in this case
Optimistic Concurrency in OrientDB
Optimistic concurrency control is used in environments with low data contention. That is, where conflicts are rare and transactions can complete without the expense of managing locks and without having transactions wait for locks to clear. This means a reduced throughput over other concurrency control methods.
OrientDB uses OCC for both Atomic Operations and Transactions.
Atomic Operations
OrientDB supports Multi-Version Concurrency Control, or MVCC, with atomic operations. This allows it to avoid locking server side resources. At the same time, it checks the version in the database. If the version is equal to the record version contained in the operation, the operation is successful. If the version found is higher than the record version contained in the operation, then another thread or user has already updated the same record. In this case, OrientDB generates an OConcurrentModificationException
exception.
Given that behavior of this kind is normal on systems that use optimistic concurrency control, developers need to write concurrency-proof code. Under this design, the application retries transactions x times before reporting the error. It does this by catching the exception, reloading the affected records and attempting to update them again. For example, consider the code for saving a document,
int maxRetries = 10;
List<ODocument> result = db.query("SELECT FROM Client WHERE id = '39w39D32d2d'");
ODocument address = result.get(0);
for (int retry = 0; retry < maxRetries; ++retry) {
try {
// LOOKUP FOR THE INVOICE VERTEX
address.field( "street", street );
address.field( "zip", zip );
address.field( "city", cityName );
address.field( "country", countryName );
address.save();
// EXIT FROM RETRY LOOP
break;
}
catch( ONeedRetryException e ) {
// IF SOMEONE UPDATES THE ADDRESS DOCUMENT
// AT THE SAME TIME, RETRY IT.
}
}
Transactions
OrientDB supports optimistic transactions. The database does not use locks when transactions are running, but when the transaction commits, each record (document or graph element) version is checked to see if there have been updates from another client. For this reason, you need to code your applications to be concurrency-proof.
Optimistic concurrency requires that you retire the transaction in the event of conflicts. For example, consider a case where you want to connect a new vertex to an existing vertex:
int maxRetries = 10;
for (int retry = 0; retry < maxRetries; ++retry) {
try {
// LOOKUP FOR THE INVOICE VERTEX
Vertex invoice = graph.getVertices("invoiceId", 2323);
// CREATE A NEW ITEM
Vertex invoiceItem = graph.addVertex("class:InvoiceItem");
invoiceItem.field("price", 1000);
// ADD IT TO THE INVOICE
invoice.addEdge(invoiceItem);
graph.commit();
// EXIT FROM RETRY LOOP
break;
}
catch( OConcurrentModificationException e ) {
// SOMEONE HAS UPDATED THE INVOICE VERTEX
// AT THE SAME TIME, RETRY IT
}
}
Concurrency Level
Before v2.2.4, transactions acquire an exclusive lock on the storage, so no matter if you have 1 or 100 cores, the execution was always serialized. With 2.2.4 and further, transactions are executed in parallel only if they involve different clusters and indexes.
In order to use the transaction parallelism, the domain has to be slightly changed by using the OrientDB inheritance by creating a base class and multiple sub-classes, one per core. Example of creating the class Log with 4 sub-classes (4 cores) and the indexed property 'id':
CREATE CLASS Log ABSTRACT
CREATE CLASS Log_1 EXTENDS Log
CREATE PROPERTY Log_1.id STRING
CREATE INDEX Log1Id ON Log_1(id) UNIQUE_HASHINDEX
CREATE CLASS Log_2 EXTENDS Log
CREATE PROPERTY Log_2.id STRING
CREATE INDEX Log1Id ON Log_2(id) UNIQUE_HASHINDEX
CREATE CLASS Log_3 EXTENDS Log
CREATE PROPERTY Log_3.id STRING
CREATE INDEX Log1Id ON Log_3(id) UNIQUE_HASHINDEX
CREATE CLASS Log_4 EXTENDS Log
CREATE PROPERTY Log_4.id STRING
CREATE INDEX Log1Id ON Log_4(id) UNIQUE_HASHINDEX
After creating multiple sub-classes, you should bind your threads/client (it depends, respectively, if you are working in embedded mode or client/server) to a different sub-class. For example with 4 cores, you have 4 sub-classes (like above) and this could be the binding for the class "Log":
- Thread/Client 1 -> Class Log_1
- Thread/Client 2 -> Class Log_2
- Thread/Client 3 -> Class Log_3
- Thread/Client 4 -> Class Log_4
If you are working with graphs, it's a good practice to apply the same rule to both vertex and edge classes. In this example we have 4 cores, so 4 clusters per vertex class and 4 clusters per edge class:
- Thread/Client 1 -> Classes User_1 and City_1 for vertices and Class Born_1 for edges
- Thread/Client 2 -> Classes User_2 and City_2 for vertices and Class Born_2 for edges
- Thread/Client 3 -> Classes User_3 and City_3 for vertices and Class Born_3 for edges
- Thread/Client 4 -> Classes User_4 and City_4 for vertices and Class Born_4 for edges
Now look at these 2 SQL scripts:
Client 1:
BEGIN
LET v1 = CREATE VERTEX User_1 SET name = 'Luca'
LET v2 = CREATE VERTEX City_1 SET name = 'Rome'
CREATE EDGE Born_1 FROM $v1 TO $v2
COMMIT RETRY 10
Client 2:
BEGIN
LET v1 = CREATE VERTEX User_2 SET name = 'Luca'
LET v2 = CREATE VERTEX City_2 SET name = 'Rome'
CREATE EDGE Born_2 FROM $v1 TO $v2
COMMIT RETRY 10
In this case the two transactions go in parallel with no conflict, because they work on different classes and indexes.
Thanks to the OrientDB polymorphism, sub-classes are instance of
the abstract class, so you can still execute queries by using the base class as target and OrientDB will consider all the sub-classes, so your model remains clean at application level. Example:
SELECT * FROM User WHERE name = 'Luca'
But if you already know that Luca if exists is in the 2nd partition of the User class (User_2 sub class), you can also execute:
SELECT * FROM User_2 WHERE name = 'Luca'
When it's possible to pre-determine there the record is saved, using the sub-class as target has better performance.
Concurrency when Adding Edges
Consider the case where multiple clients attempt to add edges on the same vertex. OrientDB could throw the OConcurrentModificationException
exception. This occurs because collections of edges are kept on vertices, meaning that, every time OrientDB adds or removes an edge, both vertices update and their versions increment. You can avoid this issue by using RIDBAG Bonsai structure, which are never embedded, so the edge never updates the vertices.
To use this configuration at run-time, before launching OrientDB, use this code:
OGlobalConfiguration.RID_BAG_EMBEDDED_TO_SBTREEBONSAI_THRESHOLD.setValue(-1);
Alternatively, you can set a parameter for the Java virtual-machine on startup, or even at run-time, before OrientDB is used:
$ java -DridBag.embeddedToSbtreeBonsaiThreshold=-1
While running in distributed mode SBTrees are not supported. If using a distributed database then you must set ridBag.embeddedToSbtreeBonsaiThreshold = Integer.MAX_VALUEto avoid replication errors. |
---|
Troubleshooting
Reduce Transaction Size
On occasion, OrientDB throws the OConcurrentModificationException
exception even when you concurrently update the first element. In particularly large transactions, where you have thousands of records involved in a transaction, one changed record is enough to roll the entire process back with an OConcurrentModificationException
exception.
To avoid issues of this kind, if you plan to update many elements in the same transaction with high-concurrency on the same vertices, a best practice is to reduce the transaction size.
Indexes
OrientDB supports five index algorithms:
- SB-Tree Index Provides a good mix of features available from other index types, good for general use. It is durable, transactional and supports range queries. It is the default index type.
- Hash Index Provides fast lookup and is very light on disk usage. It is durable and transactional, but does not support range queries. It works like a HashMap, which makes it faster on punctual lookups and it consumes less resources than other index types.
- Auto Sharding Index Provides an implementation of a DHT. It is durable and transactional, but does not support range queries. (Since v2.2)
- Lucene Full Text Index Provides good full-text indexes, but cannot be used to index other types. It is durable, transactional and supports range queries.
- Lucene Spatial Index Provides good spatial indexes, but cannot be used to index other types. It is durable, transactional and supports range queries.
Understanding Indexes
OrientDB can handle indexes in the same manner as classes, using the SQL language and prefixing the name with index:
followed by the index name. An index is like a class with two properties:
key
The index key.rid
The Record ID, which points to the record associated with the key.
Index Target
OrientDB can use two methods to update indexes:
-
Automatic Where the index is bound to schema properties. (For example,
User.id
.) If you have a schema-less database and you want to create an automatic index, then you need to create the class and the property before using the index. -
Manual Where the index is handled by the application developer, using the Java API and SQL commands (see below). You can use them as Persistent Maps, where the entry's value are the records pointed to by the index.
You can rebuild automatic indexes using the REBUILD INDEX
command.
Index Types
When you create an index, you create it as one of several available algorithm types. Once you create an index, you cannot change its type. OrientDB supports four index algorithms and several types within each. You also have the option of using any third-party index algorithms available through plugins.
- SB-Tree Algorithm
UNIQUE
These indexes do not allow duplicate keys. For composite indexes, this refers to the uniqueness of the composite keys.NOTUNIQUE
These indexes allow duplicate keys.FULLTEXT
These indexes are based on any single word of text. You can use them in queries through theCONTAINSTEXT
operator.DICTIONARY
These indexes are similar to those that useUNIQUE
, but in the case of duplicate keys, they replaces the existing record with the new record.
- HashIndex Algorithm
UNIQUE_HASH_INDEX
These indexes do not allow duplicate keys. For composite indexes, this refers to the uniqueness of the composite keys. Available since version 1.5.x.NOTUNIQUE_HASH_INDEX
These indexes allow duplicate keys. Available since version 1.5.x.FULLTEXT_HASH_INDEX
These indexes are based on any single word of text. You can use them in queries through theCONTAINSTEXT
operator. Available since version 1.5.x.DICTIONARY_HASH_INDEX
These indexes are similar to those that useUNIQUE_HASH_INDEX
, but in cases of duplicate keys, they replaces the existing record with the new record. Available since version 1.5.x.
- HashIndex Algorithm (Since v2.2)
UNIQUE_HASH_INDEX
These indexes do not allow duplicate keys. For composite indexes, this refers to the uniqueness of the composite keys.NOTUNIQUE_HASH_INDEX
These indexes allow duplicate keys.
-
- Lucene Engine
FULLTEXT
These indexes use the Lucene engine to index string content. You can use them in queries with theLUCENE
operator.SPATIAL
These indexes use the Lucene engine to index geospatial coordinates.
Every database has a default manual index type DICTIONARY
, which uses strings as keys. You may find this useful in handling the root records of trees and graphs, and handling singleton records in configurations.
Indexes and Null Values
Starting from v2.2, Indexes do not ignore NULL values, but they are indexes as any other values. This means that if you have a UNIQUE index, you cannot have multiple NULL keys. This applies only to the new indexes, opening a database with indexes previously created, will all ignore NULL by default.
To create an index that expressly ignore nulls (like the default with v2.1 and earlier), look at the following examples by using SQL or Java API.
SQL:
orientdb> CREATE INDEX addresses ON Employee (address) NOTUNIQUE METADATA {ignoreNullValues: true}
And Java API:
schema.getClass("Employee").getProperty("address").createIndex(OClass.INDEX_TYPE.NOTUNIQUE, new ODocument().field("ignoreNullValues",true));
Indexes and Composite Keys
Operations that work with indexes also work with indexes formed from composite keys. By its nature, a composite key is a collection of values, so, syntactically, it is a collection.
For example, consider a case where you have a class Book
, indexed by three fields: author
, title
and publicationYear
. You might use the following query to look up an individual book:
orientdb> SELECT FROM INDEX:books WHERE key = ["Donald Knuth", "The Art of Computer
Programming", 1968]
Alternatively, you can look for books over a range of years with the field publicationYear
:
orientdb> SELECT FROM INDEX:books WHERE key BETWEEN ["Donald Knuth", "The Art of
Computer Programming", 1960] AND ["Donald Knuth", "The Art of Computer
Programming", 2000]
Partial Match Searches
Occasionally, you may need to search an index record by several fields of its composite key. In these partial match searches, the remaining fields with undefined values can match any value in the result.
Only use composite indexes for partial match searches when the declared fields in the composite index are used from left to right. For instance, from the example above searching only title
wouldn't work with a composite index, since title
is the second value. But, you could use it when searching author
and title
.
For example, consider a case where you don't care when the books in your database were published. This allows you to use a somewhat different query, to return all books with the same author and title, but from any publication year.
orientdb> SELECT FROM INDEX:books WHERE key = ["Donald Knuth", "The Art of Computer
Programming"]
In the event that you also don't know the title of the work you want, you can further reduce it to only search all books with the same author.
orientdb> SELECT FROM INDEX:books WHERE key = ["Donald Knuth"]
Or, the equal,
orientdb> SELECT FROM INDEX:books WHERE key = "Donald Knuth"
Range Queries
Not all the indexes support range queries (check above). In the case of range queries, the field subject to the range must be the last one, (that is, the one on the far right). For example,
orientdb> SELECT FROM INDEX:books WHERE key BETWEEN ["Donald Knuth", "The Art of
Computer Programming", 1900] AND ["Donald Knuth", "The Art of Computer
Programming", 2014]
Operations against Indexes
Once you have a good understanding of the theoretical side of what indexes are and some of basic concepts that go into their use, it's time to consider the practical aspects of creating and using indexes with your application.
Creating Indexes
When you have created the relevant classes that you want to index, create the index. To create an automatic index, bound to a schema property, use the ON
section or use the name in the <class>.<property>
notation.
Syntax:
CREATE INDEX <name> [ON <class-name> (prop-names)] <type> [<key-type>]
[METADATA {<metadata>}]
-
<name>
Provides the logical name for the index. You can also use the<class.property>
notation to create an automatic index bound to a schema property. In this case, for<class>
use the class of the schema and<property>
the property created in the class.Bear in mind that this means case index names cannot contain the period (
.
) symbol, as OrientDB would interpret the text after as a property. -
<class-name>
Provides the name of the class that you are creating the automatic index to index. This class must already exist in the database. -
<prop-names>
Provides a comma-separated list of properties, which you want the automatic index to index. These properties must already exist in the schema.If the property belongs to one of the Map types, (such as
LINKMAP
, orEMBEDDEDMAP
), you can specify the keys or values to use in generating indexes. You can do this with theBY KEY
orBY VALUE
expressions, if nothing is specified, these keys are used during index creation. -
<type>
Provides the algorithm and type of index that you want to create. For information on the supported index types, see Index Types. -
<key-type>
Provides the optional key type. With automatic indexes, the key type OrientDB automatically determines the key type by reading the target schema property where the index is created. With manual indexes, if not specified, OrientDB automatically determines the key type at run-time, during the first insertion by reading the type of the class. -
<metadata>
Provides a JSON representation
Examples:
-
Creating custom indexes, deprecated since 3.0:
orientdb>
CREATE INDEX mostRecentRecords UNIQUE date
-
Creating another index for the property
id
of the classUser
:orientdb>
CREATE PROPERTY User.id BINARY
orientdb>CREATE INDEX indexForId ON User (id) UNIQUE
-
Creating indexes for property
thumbs
on classMovie
:orientdb>
CREATE INDEX thumbsAuthor ON Movie (thumbs) UNIQUE
orientdb>CREATE INDEX thumbsAuthor ON Movie (thumbs BY KEY) UNIQUE
orientdb>CREATE INDEX thumbsValue on Movie (thumbs BY VALUE) UNIQUE
-
Creating composite indexes:
orientdb>
CREATE PROPERTY Book.author STRING
orientdb>CREATE PROPERTY Book.title STRING
orientdb>CREATE PROPERTY Book.publicationYears EMBEDDEDLIST INTEGER
orientdb>CREATE INDEX books ON Book (author, title, publicationYears) UNIQUE
For more information on creating indexes, see the CREATE INDEX
command.
Dropping Indexes
In the event that you have an index that you no longer want to use, you can drop it from the database. This operation does not remove linked records.
Syntax:
DROP INDEX <name>
<name>
provides the name of the index you want to drop.
For more information on dropping indexes, see the DROP INDEX
command.
Querying Indexes
When you have an index created and in use, you can query records in the index using the SELECT
command.
Syntax:
SELECT FROM INDEX:<index-name> WHERE key = <key>
Example:
-
Selecting from the index
dictionary
where the key matches toLuke
:orientdb>
SELECT FROM INDEX:dictionary WHERE key='Luke'
Case-insensitive Matching with Indexes
In the event that you would like the index to use case-insensitive matching, set the COLLATE
attribute of the indexed properties to ci
. For instance,
orientdb> CREATE INDEX OUser.name ON OUser (name COLLATE ci) UNIQUE
Inserting Index Entries
You can insert new entries into the index using the key
and rid
pairings.
Syntax:
INSERT INTO INDEX:<index-name> (key,rid) VALUES (<key>,<rid>)
Example:
-
Inserting the key
Luke
and Record ID#10:4
into the indexdictionary
:orientdb>
INSERT INTO INDEX:dictionary (key, rid) VALUES ('Luke', #10:4)
Querying Index Ranges
In addition to querying single results from the index, you can also query a range of results between minimum and maximum values. Bear in mind that not all index types support this operation.
Syntax:
SELECT FROM INDEX:<index-name> WHERE key BETWEEN <min> AND <max>
Example:
-
Querying from the index
coordinates
and range between10.3
and10.7
:orientdb>
SELECT FROM INDEX:coordinates WHERE key BETWEEN 10.3 AND 10.7
Removing Index Entries
You can delete entries by passing the key
and rid
values. This operation returns TRUE
if the removal was successful and FALSE
if the entry wasn't found.
Syntax:
DELETE FROM INDEX:<index-name> WHERE key = <key> AND rid = <rid>
Example:
-
Removing an entry from the index
dictionary
:orientdb>
DELETE FROM INDEX:dictionary WHERE key = 'Luke' AND rid = #10:4
Removing Index Entries by Key
You can delete all entries from the index through the requested key.
Syntax:
DELETE FROM INDEX:<index-name> WHERE key = <key>
Example:
-
Delete entries from the index
addressbook
whee the key matches toLuke
:orientdb>
DELETE FROM INDEX:addressbook WHERE key = 'Luke'
Removing Index Entries by RID
You can remove all index entries to a particular record by its record ID.
Syntax:
DELETE FROM INDEX:<index-name> WHERE rid = <rid>
Example:
-
Removing entries from index
dictionary
tied to the record ID#10:4
:orientdb>
DELETE FROM INDEX:dictionary WHERE rid = #10:4
Counting Index Entries
To see the number of entries in a given index, you can use the COUNT()
function.
Syntax:
SELECT COUNT(*) AS size FROM INDEX:<index-name>
Example:
-
Counting the entries on the index
dictionary
:orientdb>
SELECT COUNT(*) AS size FROM INDEX:dictionary
Querying Keys from Indexes
You can query all keys in an index using the SELECT
command.
Syntax:
SELECT key FROM INDEX:<index-name>
Example:
-
Querying the keys in the index
dictionary
:orientdb>
SELECT key FROM INDEX:dictionary
Querying Index Entries
You can query for all entries on an index as key
and rid
pairs.
Syntax:
SELECT key, value FROM INDEX:<index-name>
Example:
-
Querying the
key
/rid
pairs from the indexdictionary
:orientdb>
SELECT key, value FROM INDEX:dictionary
Clearing Indexes
Remove all entries from an index. After running this command, the index is empty.
Syntax:
DELETE FROM INDEX:<index-name>
Example:
-
Removing all entries from the index
dictionary
:orientdb>
DELETE FROM INDEX:dictionary
Query the available indexes
To access to the indexes, you can use SQL.
Create your index engine
Here you can find a guide how to create a custom index engine.
SB-Tree Index Algorithm
This indexing algorithm provides a good mix of features, similar to the features available from other index types. It is good for general use and is durable, transactional and supports range queries. There are four index types that utilize the SB-Tree index algorithm:
UNIQUE
Does not allow duplicate keys, fails when it encounters duplicates.NOTUNIQUE
Does allow duplicate keys.FULLTEXT
Indexes to any single word of text.DICTIONARY
Does not allow duplicate keys, overwrites when it encounters duplicates.
For more information on
FULLTEXT_HASH_INDEX
, see FullText Index.
The SB-Tree index algorithm is based on the B-Tree index algorithm. It has been adapted with several optimizations, which relate to data insertion and range queries. As is the case with all other tree-based indexes, SB-Tree index algorithm experiences log(N)
complexity, but the base to this logarithm is about 500.
NOTE: There is an issue in the replacement of indexes based on B-Tree with those based on COLA Tree to avoid slowdowns introduced by random I/O operations. For more information see Issue #1756.
Hash Index Algorithm
This indexing algorithm provides a fast lookup and is very light on disk usage. It is durable and transactional, but does not support range queries. It is similar to a HashMap, which makes it faster on punctual lookups and it consumes less resources than other index types. The Hash index algorithm supports four index types, which have been available since version 1.5.x:
UNIQUE_HASH_INDEX
Does not allow duplicate keys, it fails when it encounters duplicates.NOTUNIQUE_HASH_INDEX
Does allow duplicate keys.FULLTEXT_HASH_INDEX
Indexes to any single word.DICTIONARY
Does not allow duplicate keys, it overwrites when it encounters duplicates.
For more information on
FULLTEXT_HASH_INDEX
, see FullText Index.
Hash indexes are able to perform index read operations in one I/O operation and write operations in a maximum of three I/O operations. The Hash Index algorithm is based on the Extendible Hashing algorithm. Despite not providing support for range queries, it is noticeably faster than SB-Tree Index Algorithms, (about twice as fast when querying through ten million records).
NOTE: There is an issue relating to the enhancement of Hash indexes to avoid slowdowns introduced by random I/O operations using LSM Tree approaches. For more information, see Issue #1757.
Auto Sharding Index Algorithm
(Since v2.2)
This indexing algorithm is based on the DHT concept, where they keys are stored on different partition, based on the Murmur3 hash function.
Auto Sharding Index supports the following index types:
UNIQUE_HASH_INDEX
Does not allow duplicate keys, it fails when it encounters duplicates.NOTUNIQUE_HASH_INDEX
Does allow duplicate keys.
Under the hood, this index creates multiple Hash Indexes, one per cluster. So if you have 8 clusters for the class "Employee", this index will create, at the beginning, 8 Hash Indexes.
Since this index is based on the Hash Index, it's able to perform index read operations in one I/O operation and write operations in a maximum of three I/O operations. The Hash Index algorithm is based on the Extendible Hashing algorithm. Despite not providing support for range queries, it is noticeably faster than SB-Tree Index Algorithms, (about twice as fast when querying through ten million records).
Usage
Create an index by passing "AUTOSHARDING" as index engine:
final OClass cls = db.createClass("Log");
cls.createProperty("key", OType.LONG);
cls.createIndex("idx_LogKey", OClass.INDEX_TYPE.UNIQUE.toString(),
(OProgressListener) null, (ODocument) null, "AUTOSHARDING", new String[] { "key" });
Performance
On multi-core hw, using this index instead of Hash Index gives about +50% more throughput on insertion on a 8 cores machine.
Distributed
The fully distributed version of this index will be supported in v3.0. In v2.2 each node has own copy of the index with all the partitions.
Internals
This is the algorithm for the put(key,value)
:
int partition = Murmur3_hash(key) % partitions;
getSubIndex(partition).put(key,value);
This is for the value = get(key)
:
int partition = Murmur3_hash(key) % partitions;
return getSubIndex(partition).get(key);
FullText Indexes
The SB-Tree index algorithm provides support for FullText indexes. These indexes allow you to index text as a single word and its radix. FullText indexes are like having a search engine on your database.
NOTE: Bear in mind that there is a difference between
FULLTEXT
without theLUCENE
operator, which uses a FullText index with the SB-Tree index algorithm andFULLTEXT
with theLUCENE
operator, which uses a FullText index through the Lucene Engine.For more information on the latter, see Lucene FullText Index.
Creating FullText Indexes
If you want to create an index using the FullText SB-Tree index algorithm, you can do so using the CREATE INDEX
command.
orientdb> CREATE INDEX City.name ON City(name) FULLTEXT
This creates a FullText index on the property name
of the class City
, using the default configuration.
FullText Index Parameters
In the event that the default FullText Index configuration is not sufficient to your needs, there are a number of parameters available to fine tune how it generates the index.
Parameter | Default | Description |
---|---|---|
indexRadix | TRUE | Word prefixes will be also index |
ignoreChars | " | Chars to skip when indexing |
separatorChars | \r\n\t:;,.|+*/\=!?[](.md) | |
minWordLength | 3 | Minimum word length to index |
stopWords | the in a at as and or for his her him this that what which while up with be was were is | Stop words excluded from indexing |
To configure a FullText Index, from version 1.7 on, you can do so through the OrientDB console or the Java API. When configuring the index from the console, use the CREATE INDEX
command with the METADATA
operator.
orientdb> CREATE INDEX City.name ON City(name) FULLTEXT METADATA
{"indexRadix": true, "ignoreChars": "&", "separatorChars": " |()",
"minWordLength": 4, "stopWords": ["the", "of"]}
Alternatively, you can configure the index in Java.
OClass city = db.getClass("City");
ODocument metadata = new ODocument();
metadata.field("indexRadix", true);
metadata.field("stopWords", Arrays.asList(new String[] { "the", "in", "a", "at" }));
metadata.field("separatorChars", " :;?[](.md)");
metadata.field("ignoreChars", "$&");
metadata.field("minWordLength", 5);
city.createIndex("City.name", "FULLTEXT", null, metadata, null, new String[] { "name" });
Lucene FullText Index
In addition to the standard FullText Index, which uses the SB-Tree index algorithm, you can also create FullText indexes using the Lucene Engine . Apache LuceneTM is a high-performance, full-featured text search engine library written entirely in Java. Check the Lucene documentation for a full overview of its capabilities.
How Lucene's works?
Let's look at a sample corpus of five documents:
- My sister is coming for the holidays.
- The holidays are a chance for family meeting.
- Who did your sister meet?
- It takes an hour to make fudge.
- My sister makes awesome fudge.
What does Lucene do? Lucene is a full text search library. Search has two principal stages: indexing and retrieval.
During indexing, each document is broken into words, and the list of documents containing each word is stored in a list called the postings list. The posting list for the word my is:
my --> 1,5
Posting list for others terms:
fudge --> 4,5
sister --> 1,2,3,5
fudge --> 4,5
The index consists of all the posting lists for the words in the corpus. Indexing must be done before retrieval, and we can only retrieve documents that were indexed.
Retrieval is the process starting with a query and ending with a ranked list of documents. Say the query is "my fudge". In order to find matches for the query, we break it into the individual words, and go to the posting lists. The full list of documents containing the keywords is [1,4,5]. Note that the query is broken into words (terms) and each term is matched with the terms in the index. Lucene's default operator is OR, so it retrieves the documents tha contain my OR fudge. If we want to retrieve documents that contain both my and fudge, rewrite the query: "+my +fudge".
Lucene doesn't work as a LIKE operator on steroids, it works on single terms. Terms are produced analyzing the provided text, so the right analyzer should be configured. On the other side, it offers a complete query language, well documented here:
Index creation
To create an index based on Lucene
CREATE INDEX ON (prop-names) FULLTEXT ENGINE LUCENE [{json metadata}]
The following SQL statement will create a FullText index on the property name
for the class City
, using the Lucene Engine.
CREATE INDEX City.name ON City(name) FULLTEXT ENGINE LUCENE
Indexes can also be created on n-properties. For example, create an index on the properties name
and description
on the class City
.
CREATE INDEX City.name_description ON City(name, description)
FULLTEXT ENGINE LUCENE
When multiple properties should be indexed, define a single multi-field index over the class. A single multi-field index needs less resources, such as file handlers. Moreover, it is easy to write better Lucene queries. The default analyzer used by OrientDB when a Lucene index is created is the StandardAnalyzer. The StandardAnalyzer usually works fine with western languages, but Lucene offers analyzer for different languages and use cases.
Two minutes tutorial
Open studio or console and create a sample dataset:
CREATE CLASS Item;
CREATE PROPERTY Item.text STRING;
CREATE INDEX Item.text ON Item(text) FULLTEXT ENGINE LUCENE;
INSERT INTO Item (text) VALUES ('My sister is coming for the holidays.');
INSERT INTO Item (text) VALUES ('The holidays are a chance for family meeting.');
INSERT INTO Item (text) VALUES ('Who did your sister meet?');
INSERT INTO Item (text) VALUES ('It takes an hour to make fudge.');
INSERT INTO Item (text) VALUES ('My sister makes awesome fudge.');
Search all documents that contain sister:
SELECT FROM Item WHERE SEARCH_CLASS("sister") = true
Search all documents that contain sister AND coming:
SELECT FROM Item WHERE SEARCH_CLASS("+sister +coming") = true
Search all documents that contain sister but NOT coming:
SELECT FROM Item WHERE SEARCH_CLASS("+sister -coming") = true
Search all documents that contain the phrase sister meet:
SELECT FROM Item WHERE SEARCH_CLASS(' "sister meet" ') = true
Search all documents that contain terms starting with meet:
SELECT FROM Item WHERE SEARCH_CLASS('meet*') = true
To better understand how the query parser work, read carefully the official documentation and play with the above documents.
Customize Analyzers
In addition to the StandardAnalyzer, full text indexes can be configured to use different analyzer by the METADATA
operator through CREATE INDEX
.
Configure the index on City.name
to use the EnglishAnalyzer
:
CREATE INDEX City.name ON City(name)
FULLTEXT ENGINE LUCENE METADATA {
"analyzer": "org.apache.lucene.analysis.en.EnglishAnalyzer"
}
Configure the index on City.name
to use different analyzers for indexing and querying.
CREATE INDEX City.name ON City(name)
FULLTEXT ENGINE LUCENE METADATA {
"index": "org.apache.lucene.analysis.en.EnglishAnalyzer",
"query": "org.apache.lucene.analysis.standard.StandardAnalyzer"
}
EnglishAnalyzer
will be used to analyze text while indexing and the StandardAnalyzer
will be used to analyze query text.
A very detailed configuration, on multi-field index configuration, could be:
CREATE INDEX Song.fulltext ON Song(name, lyrics, title, author, description)
FULLTEXT ENGINE LUCENE METADATA {
"default": "org.apache.lucene.analysis.standard.StandardAnalyzer",
"index": "org.apache.lucene.analysis.core.KeywordAnalyzer",
"query": "org.apache.lucene.analysis.standard.StandardAnalyzer",
"name_index": "org.apache.lucene.analysis.standard.StandardAnalyzer",
"name_query": "org.apache.lucene.analysis.core.KeywordAnalyzer",
"lyrics_index": "org.apache.lucene.analysis.en.EnglishAnalyzer",
"title_index": "org.apache.lucene.analysis.en.EnglishAnalyzer",
"title_query": "org.apache.lucene.analysis.en.EnglishAnalyzer",
"author_query": "org.apache.lucene.analysis.core.KeywordAnalyzer",
"description_index": "org.apache.lucene.analysis.standard.StandardAnalyzer",
"description_index_stopwords": [
"the",
"is"
]
}
With this configuration, the underlying Lucene index will works in different way on each field:
- name: indexed with StandardAnalyzer, searched with KeywordAnalyzer (it's a strange choice, but possible)
- lyrics: indexed with EnglishAnalyzer, searched with default query analyzer StandardAnalyzer
- title: indexed and searched with EnglishAnalyzer
- author: indexed and searched with KeywordAnalyzer
- description: indexed with StandardAnalyzer with a given set of stop-words that overrides the internal set
Analysis is the foundation of Lucene. By default the StandardAnalyzer removes english stop-words and punctuation and lowercase the generated terms:
The holidays are a chance for family meeting!
Would produce
- holidays
- are
- chance
- for
- family
- meeting
Each analyzer has its set of stop-words and tokenize the text in a different way. Read the full (documentation)[http://lucene.apache.org/core/6_6_0/].
Query parser
It is possible to configure some behavior of the Lucene query parser Query parser's behavior can be configured at index creation time and overridden at runtime.
Allow Leading Wildcard
Lucene by default doesn't support leading wildcard: Lucene wildcard support
It is possible to override this behavior with a dedicated flag on meta-data:
{
"allowLeadingWildcard": true
}
CREATE INDEX City.name ON City(name)
FULLTEXT ENGINE LUCENE METADATA {
"allowLeadingWildcard": true
}
Use this flag carefully, as stated in the Lucene FAQ:
Note that this can be an expensive operation: it requires scanning the list of tokens in the index in its entirety to look for those that match the pattern.
Disable lower case on terms
Lucene's QueryParser applies a lower case filter on expanded queries by default. It is possible to override this behavior with a dedicated flag on meta-data:
{
"lowercaseExpandedTerms": false
}
It is useful when used in pair with keyword analyzer:
CREATE INDEX City.name ON City(name)
FULLTEXT ENGINE LUCENE METADATA {
"lowercaseExpandedTerms": false,
"default" : "org.apache.lucene.analysis.core.KeywordAnalyzer"
}
With lowercaseExpandedTerms set to false, these two queries will return different results:
SELECT from Person WHERE SEARCH_CLASS("NAME") = true
SELECT from Person WHERE WHERE SEARCH_CLASS("name") = true
Querying Lucene FullText Indexes
OrientDB 3.0.x introduced search functions: SEARCH_CLASS, SEARCH_FIELDS, SEARCH_INDEX, SEARCH_MORE Every function accepts as last, optional, parameter a JSON with additional configuration.
SEARCH_CLASS
The best way to use the search capabilities of OrientDB is to define a single multi-fields index and use the SEARCH_CLASS function. In case more than one full-text index are defined over a class, an error is raised in case of SEARCH_CLASS invocation.
Suppose to have this index
CREATE INDEX City.fulltext ON City(name, description) FULLTEXT ENGINE LUCENE
A query that retrieve cities with the name starting with cas and description containing the word beautiful:
SELECT FROM City WHERE SEARCH_CLASS("+name:cas* +description:beautiful") = true
The function accepts metadata JSON as second parameter:
SELECT FROM City WHERE SEARCH_CLASS("+name:cas* +description:beautiful", {
"allowLeadingWildcard": true ,
"lowercaseExpandedTerms": false,
"boost": {
"name": 2
},
"highlight": {
"fields": ["name"],
"start": "",
"end": ""
}
}) = true
The query shows query parser's configuration overrides, boost of field name with highlight. Highlight and boost will be explained later.
SEARCH_MORE
OrientDB exposes the Lucene's more like this capability with a dedicated function.
The first parameter is the array of RID of elements to be used to calculate similarity, the second parameter the usual metadata JSON used to tune the query behaviour.
SELECT FROM City WHERE SEARCH_MORE([#25:2, #25:3],{'minTermFreq':1, 'minDocFreq':1} ) = true
It is possible to use a query to gather RID of documents to be used to calculate similarity:
SELECT FROM City
let $a=(SELECT @rid FROM City WHERE name = 'Rome')
WHERE SEARCH_MORE( $a, { 'minTermFreq':1, 'minDocFreq':1} ) = true
Lucene's MLT has a lot of parameter, and all these are exposed through the metadata JSON: http://lucene.apache.org/core/6_6_0/queries/org/apache/lucene/queries/mlt/MoreLikeThis.html
- fieldNames: array of field's names to be used to extract content
- maxQueryTerms
- minDocFreq
- maxDocFreq
- minTermFreq
- boost
- boostFactor
- maxWordLen
- minWordLen
- maxNumTokensParsed
- stopWords
Query parser's runtime configuration
It is possible to override the query parser's configuration given at creation index time at runtime passing a json:
SELECT from Person WHERE SEARCH_CLASS("bob",{
"allowLeadingWildcard": true ,
"lowercaseExpandedTerms": false
} ) = true
The same can be done for query analyzer, overriding the configuration given at index creation's time:
SELECT from Person WHERE SEARCH_CLASS("bob",{
"customAnalysis": true ,
"query": "org.apache.lucene.analysis.standard.StandardAnalyzer",
"name_query": "org.apache.lucene.analysis.en.EnglishAnalyzer"
} ) = true
The customAnalysis flag is mandatory to enable the runtime configuration of query analyzers. The runtime configuration is per query and it isn't stored nor reused for a subsequent query. The custom configuration can be used with all the functions.
SEARCH_INDEX
The SEARCH_INDEX function allows to execute the query on a single index. It is useful if more than one index are defined over a class.
SELECT FROM City WHERE SEARCH_INDEX("City.name", "cas*") = true
The function accepts a JSON as third parameter, as for SEARCH_CLASS.
SEARCH_FIELDS
The SEARCH_FIELDS function allows to execute query over the index that is defined over one or more fields:
SELECT FROM City WHERE SEARCH_FIELDS(["name", "description"], "name:cas* description:beautiful") = true
The function accepts a JSON as third parameter, as for SEARCH_CLASS.
Numeric and date range queries
If the index is defined over a numeric field (INTEGER, LONG, DOUBLE) or a date field (DATE, DATETIME), the engine supports range queries
Suppose to have a City
class with a multi-field Lucene index defined:
CREATE CLASS CITY EXTENDS V
CREATE PROPERTY CITY.name STRING
CREATE PROPERTY CITY.size INTEGER
CREATE INDEX City.name ON City(name,size) FULLTEXT ENGINE LUCENE
Then query using ranges:
SELECT FROM City WHERE SEARCH_CLASS('name:cas* AND size:[15000 TO 20000]') = true
Ranges can be applied to DATE/DATETIME field as well. Create a Lucene index over a property:
CREATE CLASS Article EXTENDS V
CREATE PROPERTY Article.createdAt DATETIME
CREATE INDEX Article.createdAt ON Article(createdAt) FULLTEXT ENGINE LUCENE
Then query to retrieve articles published only in a given time range:
SELECT FROM Article WHERE SEARCH_CLASS('[201612221000 TO 201612221100]') =true
Retrieve the Score
When the lucene index is used in a query, the results set carries a context variable for each record representing the score.
To display the score add $score
in projections.
SELECT *,$score FROM V WHERE name LUCENE "test*"
Highlighting
OrientDB uses the Lucene's highlighter. Highlighting can be configured using the metadata JSON. The highlighted content of a field is returned in a dedicated field suffixed with _hl:
SELECT name, $name_hl, description, $description_hl FROM City
WHERE SEARCH_CLASS("+name:cas* +description:beautiful", {
"highlight": {
"fields": ["name", "description"],
"start": "",
"end": ""
}
}) = true
Parameters
- fields: array of field names to be highlighted
- start: start delimiter for highlighted text (default <B>)
- end: end delimiter for highlighted text (default </B>)
- maxNumFragments: maximum number of text's fragments to highlight
Sorting
Sorting uses SortedDocValuesField which for too large string can cause DocValuesField is too large, must be <= 32766
exception.
In order to disable sorting add the configuration on the metadata ad index creation (From OrientDB 3.0.24).
For all fields:
CREATE INDEX City.name ON City(name)
FULLTEXT ENGINE LUCENE METADATA {
"*_index_sorted": false
}
For single field:
CREATE INDEX City.name ON City(name)
FULLTEXT ENGINE LUCENE METADATA {
"name_index_sorted": false
}
Documents retrieved by a search call are ordered by their score. It is possible to configure the way the document are sorted. Read carefully the official documentation about sorting : https://lucene.apache.org/core/6_6_1/core/org/apache/lucene/search/Sort.html
SELECT name, description, size FROM City
WHERE SEARCH_CLASS("+name:cas* +description:beautiful", {
"sort": [ { 'field': 'size', reverse:true, type:'INT' }]
}) = true
Sort over multiple fields is possible:
SELECT name, description, size FROM City
WHERE SEARCH_CLASS("+name:cas* +description:beautiful", {
"sort": [
{ 'field': 'size', reverse:true, type:'INT' },
{ 'field': 'name', reverse:false, type:'STRING' },
{ reverse:false, type:'DOC' },
]
}) = true
Sort configuration:
- field: is the field name. Could be absent only if the sort type is DOC or INDEX
- reverse: if set to true, will sort for the given field in reverse order
- type: look to https://lucene.apache.org/core/6_6_1/core/org/apache/lucene/search/SortField.Type.html
CUSTOM type is not supported
Cross class search (Enterprise Edition)
Bundled with the enterprise edition there's the SEARCH_CROSS function that is able to search over all the Lucene indexes defined on a database
Suppose to define two indexes:
CREATE INDEX Song.title ON Song (title,author) FULLTEXT ENGINE LUCENE METADATA
CREATE INDEX Author.name on Author(name,score) FULLTEXT ENGINE LUCENE METADATA
Searching for a term on each class implies a lot of different queries to be aggregated.
The SEARCH_CLASS function automatically performs the given query on each full-text index configured inside the database.
SELECT EXPAND(SEARCH_CROSS('beautiful'))
The query will be execute over all the indexes configured on each field. It is possible to search over a given field of a certain class, just qualify the field names with their class name:
SELECT EXPAND(SEARCH_CROSS('Song.title:beautiful Author.name:bob'))
Another way is to use the metadata field _CLASS present in every index:
SELECT expand(SEARCH_CROSS('(+_CLASS:Song +title:beautiful) (+_CLASS:Author +name:bob)') )
All the options of a Lucene's query are allowed: inline boosting, phrase queries, proximity etc.
The function accepts a metadata JSON as second parameter:
SELECT EXPAND(SEARCH_CROSS('Author.name:bob Song.title:*tain', {"
"allowLeadingWildcard" : true,
"boost": {
"Author.name": 2.0
}
}
)
Highlight isn't supported yet.
Lucene Writer fine tuning (expert)
It is possible to fine tune the behaviour of the underlying Lucene's IndexWriter
CREATE INDEX City.name ON City(name)
FULLTEXT ENGINE LUCENE METADATA {
"directory_type": "nio",
"use_compound_file": false,
"ram_buffer_MB": "16",
"max_buffered_docs": "-1",
"max_buffered_delete_terms": "-1",
"ram_per_thread_MB": "1024",
"default": "org.apache.lucene.analysis.standard.StandardAnalyzer"
}
- directory_type: configure the access type to the Lucene's index
- nio (default): the index is opened with NIOFSDirectory
- mmap: the index is opened with MMapDirectory
- ram: index will be created in memory with RAMDirectory
- use_compound_file: default is false
- ram_buffer_MB: size of the document's buffer in MB, default value is 16 MB (which means flush when buffered docs consume approximately 16 MB RAM)
- max_buffered_docs: size of the document's buffer in number of docs, disabled by default (because IndexWriter flushes by RAM usage by default)
- max_buffered_delete_terms: disabled by default (because IndexWriter flushes by RAM usage by default).
- ram_per_thread_MB: default value is 1945
For a detailed explanation of config parameters and IndexWriter behaviour
- indexWriterConfig : https://lucene.apache.org/core/6_6_0/core/org/apache/lucene/index/IndexWriterConfig.html
- indexWriter: https://lucene.apache.org/core/6_6_0/core/org/apache/lucene/index/IndexWriter.html
Index lifecycle
Lucene indexes are lazy. If the index is in idle mode, no reads and no writes, it will be closed. Intervals are fully configurable.
- flushIndexInterval: flushing index interval in milliseconds, default to 20000 (10s)
- closeAfterInterval: closing index interval in milliseconds, default to 120000 (12m)
- firstFlushAfter: first flush time in milliseconds, default to 10000 (10s)
To configure the index lifecycle, just pass the parameters in the JSON of metadata:
CREATE INDEX City.name ON City(name) FULLTEXT ENGINE LUCENE METADATA
{
"flushIndexInterval": 200000,
"closeAfterInterval": 200000,
"firstFlushAfter": 20000
}
Create index using the Java API
The FullText Index with the Lucene Engine is configurable through the Java API.
OSchema schema = databaseDocumentTx.getMetadata().getSchema();
OClass oClass = schema.createClass("Foo");
oClass.createProperty("name", OType.STRING);
oClass.createIndex("City.name", "FULLTEXT", null, null, "LUCENE", new String[] { "name"});
The LUCENE operator (deprecated)
NOTE: LUCENE operator is translated to SEARCH_FIELDS function, but it doesn't support the metadata JSON
You can query the Lucene FullText Index using the custom operator LUCENE
with the Query Parser Syntax from the Lucene Engine.
SELECT FROM V WHERE name LUCENE "test*"
This query searches for test
, tests
, tester
, and so on from the property name
of the class V
.
The query can use proximity operator ~, the required (+) and prohibit (-) operators, phrase queries, regexp queries:
SELECT FROM Article WHERE content LUCENE "(+graph -rdbms) AND +cloud"
Working with multiple fields (deprecated)
NOTE: define a single Lucene index on the class and use SEARCH_CLASS function
In addition to the standard Lucene query above, you can also query multiple fields. For example,
SELECT FROM Class WHERE [prop1, prop2] LUCENE "query"
In this case, if the word query
is a plain string, the engine parses the query using MultiFieldQueryParser on each indexed field.
To execute a more complex query on each field, surround your query with parentheses, which causes the query to address specific fields.
SELECT FROM Article WHERE [content, author] LUCENE "(content:graph AND author:john)"
Here, the engine parses the query using the QueryParser
Creating a Manual Lucene Index (deprecated)
NOTE: avoid manual Lucene index
The Lucene Engine supports index creation without the need for a class.
Syntax:
CREATE INDEX FULLTEXT ENGINE LUCENE [] [METADATA {}]
For example, create a manual index using the CREATE INDEX
command:
CREATE INDEX Manual FULLTEXT ENGINE LUCENE STRING, STRING
Once you have created the index Manual
, you can insert values in index using the INSERT INTO INDEX:...
command.
INSERT INTO INDEX:Manual (key, rid) VALUES(['Enrico', 'Rome'], #5:0)
You can then query the index through SELECT...FROM INDEX:
:
SELECT FROM INDEX:Manual WHERE key LUCENE "Enrico"
Manual indexes could be created programmatically using the Java API
ODocument meta = new ODocument().field("analyzer", StandardAnalyzer.class.getName());
OIndex> index = databaseDocumentTx.getMetadata().getIndexManager()
.createIndex("apiManual", OClass.INDEX_TYPE.FULLTEXT.toString(),
new OSimpleKeyIndexDefinition(1, OType.STRING, OType.STRING), null, null, meta, OLuceneIndexFactory.LUCENE_ALGORITHM);
Lucene Spatial
(Versions 2.2 and after only, otherwise look at Legacy section)
Since v 3.0, this module is provided in-bundle with the main distribution (Community and Enterprise Editions).
Install
In versions prior to v 3.0 the spatial plugin was a separate component and needed to be installed manually. This is not the case in v 3.0, where the spatial plugin is included in the main distribution, so there is no need to install it.
Geometry Data
OrientDB supports the following Geometry objects :
- Point (OPoint)
- Line (OLine)
- Polygon (OPolygon)
- MultiPoint (OMultiPoint)
- MultiLine (OMultiline)
- MultiPolygon (OMultiPolygon)
- Geometry Collections
OrientDB stores those objects like embedded documents with special classes. The module creates abstract classes that represent each Geometry object type, and those classes can be embedded in user defined classes to provide geospatial information.
Each spatial classes (Geometry Collection excluded) comes with field coordinates that will be used to store the geometry structure. The "coordinates" field of a geometry object is composed of one position (Point), an array of positions (LineString or MultiPoint), an array of arrays of positions (Polygons, MultiLineStrings) or a multidimensional array of positions (MultiPolygon).
Geometry data Example
Restaurants Domain
CREATE class Restaurant
CREATE PROPERTY Restaurant.name STRING
CREATE PROPERTY Restaurant.location EMBEDDED OPoint
To insert restaurants with location
From SQL
INSERT INTO Restaurant SET name = 'Dar Poeta', location = {"@class": "OPoint","coordinates" : [12.4684635,41.8914114]}
or as an alternative, if you use WKT format you can use the function ST_GeomFromText
to create the OrientDB geometry object.
INSERT INTO Restaurant SET name = 'Dar Poeta', location = St_GeomFromText("POINT (12.4684635 41.8914114)")
From JAVA
ODocument location = new ODocument("OPoint");
location.field("coordinates", Arrays.asList(12.4684635, 41.8914114));
ODocument doc = new ODocument("Restaurant");
doc.field("name","Dar Poeta");
doc.field("location",location);
doc.save();
A spatial index on the location field s defined by
CREATE INDEX Restaurant.location ON Restaurant(location) SPATIAL ENGINE LUCENE"
Functions
OrientDB follows The Open Geospatial Consortium OGC for extending SQL to support spatial data. OrientDB implements a subset of SQL-MM functions with ST prefix (Spatial Type)
ST_AsText
Syntax : ST_AsText(geom)
Example
SELECT ST_AsText({"@class": "OPoint","coordinates" : [12.4684635,41.8914114]})
Result
----------
POINT (12.4684635 41.8914114)
ST_GeomFromText
Syntax : ST_GeomFromText(text)
Example
select ST_GeomFromText("POINT (12.4684635 41.8914114)")
Result
----------------------------------------------------------------------------------
{"@type":"d","@version":0,"@class":"OPoint","coordinates":[12.4684635,41.8914114]}
ST_AsGeoJSON
Syntax : ST_AsGeoJSON(geom)
Example
select ST_AsGeoJSON(ST_GeomFromText("POINT (12.4684635 41.8914114)"))
Result
----------
{"type":"Point","coordinates":[12.468464,41.891411]}
ST_GeomFromGeoJSON
Syntax : ST_GeomFromGeoJSON(getJsonText)
Example
select ST_GeomFromGeoJSON('{"type":"Point","coordinates":[12.4684635,41.8914114]}')
Result
----------------------------------------------------------------------------------
{"@type":"d","@version":0,"@class":"OPoint","coordinates":[12.4684635,41.8914114]}
### ST_Equals
Returns true if geom1 is spatially equal to geom2
Syntax : ST_Equals(geom1,geom2)
Example
```SQL
SELECT ST_Equals(ST_GeomFromText('LINESTRING(0 0, 10 10)'), ST_GeomFromText('LINESTRING(0 0, 5 5, 10 10)'))
Result
-----------
true
ST_Within
Returns true if geom1 is inside geom2
Syntax : ST_Within(geom1,geom2)
This function will use an index if available.
Example
select * from City where ST_WITHIN(location,'POLYGON ((12.314015 41.8262816, 12.314015 41.963125, 12.6605063 41.963125, 12.6605063 41.8262816, 12.314015 41.8262816))') = true
ST_DWithin
Returns true if the geometries are within the specified distance of one another
Syntax : ST_DWithin(geom1,geom2,distance)
Example
SELECT ST_DWithin(ST_GeomFromText('POLYGON((0 0, 10 0, 10 5, 0 5, 0 0))'), ST_GeomFromText('POLYGON((12 0, 14 0, 14 6, 12 6, 12 0))'), 2.0d) as distance
SELECT from Polygon where ST_DWithin(geometry, ST_GeomFromText('POLYGON((12 0, 14 0, 14 6, 12 6, 12 0))'), 2.0) = true
ST_Contains
Returns true if geom1 contains geom2
Syntax : ST_Contains(geom1,geom2)
This function will use an index if available.
Example
SELECT ST_Contains(ST_Buffer(ST_GeomFromText('POINT(0 0)'),10),ST_GeomFromText('POINT(0 0)'))
Result
----------
true
SELECT ST_Contains(ST_Buffer(ST_GeomFromText('POINT(0 0)'),10),ST_Buffer(ST_GeomFromText('POINT(0 0)'),20))
Result
----------
false
ST_Disjoint
Returns true if geom1 does not spatially intersects geom2
Syntax: St_Disjoint(geom1,geom2)
This function does not use indexes
Example
SELECT ST_Disjoint(ST_GeomFromText('POINT(0 0)'), ST_GeomFromText('LINESTRING ( 2 0, 0 2 )'));
Result
-----------------
true
SELECT ST_Disjoint(ST_GeomFromText('POINT(0 0)'), ST_GeomFromText('LINESTRING ( 0 0, 0 2 )'));
Result
-----------------
false
ST_Intersects
Returns true if geom1 spatially intersects geom2
Syntax: ST_Intersects(geom1,geom2)
Example
SELECT ST_Intersects(ST_GeomFromText('POINT(0 0)'), ST_GeomFromText('LINESTRING ( 2 0, 0 2 )'));
Result
-------------
false
SELECT ST_Intersects(ST_GeomFromText('POINT(0 0)'), ST_GeomFromText('LINESTRING ( 0 0, 0 2 )'));
Result
-------------
true
ST_AsBinary
Returns the Well-Known Binary (WKB) representation of the geometry
Syntax : ST_AsBinary(geometry)
Example
SELECT ST_AsBinary(ST_GeomFromText('POINT(0 0)'))
ST_Envelope
Returns a geometry representing the bounding box of the supplied geometry
Syntax : ST_Envelope(geometry)
Example
SELECT ST_AsText(ST_Envelope(ST_GeomFromText('POINT(1 3)')));
Result
----------
POINT (1 3)
SELECT ST_AsText(ST_Envelope(ST_GeomFromText('LINESTRING(0 0, 1 3)')))
Result
-----------------------------------
POLYGON ((0 0, 0 3, 1 3, 1 0, 0 0))
ST_Buffer
Returns a geometry that represents all points whose distance from this Geometry is less than or equal to distance.
Syntax: ST_Buffer(geometry,distance [,config])
where config is an additional parameter (JSON) that can be use to set:
quadSegs: int -> number of segments used to approximate a quarter circle (defaults to 8).
{
quadSegs : 1
}
endCap : round|flat|square -> endcap style (defaults to "round").
{
endCap : 'square'
}
join : round|mitre|bevel -> join style (defaults to "round")
{
join : 'bevel'
}
mitre : double -> mitre ratio limit (only affects mitered join style).
{
join : 'mitre',
mitre : 5.0
}
Example
SELECT ST_AsText(ST_Buffer(ST_GeomFromText('POINT(100 90)'),50))
SELECT ST_AsText(ST_Buffer(ST_GeomFromText('POINT(100 90)'), 50, { quadSegs : 2 }));
Operators
A && B
Overlaps operator. Returns true if bounding box of A overlaps bounding box of B. This operator will use an index if available.
Example
CREATE CLASS TestLineString
CREATE PROPERTY TestLineString.location EMBEDDED OLineString
INSERT INTO TestLineSTring SET name = 'Test1' , location = St_GeomFromText("LINESTRING(0 0, 3 3)")
INSERT INTO TestLineSTring SET name = 'Test2' , location = St_GeomFromText("LINESTRING(0 1, 0 5)")
SELECT FROM TestLineString WHERE location && "LINESTRING(1 2, 4 6)"
Spatial Indexes
To speed up spatial search and match condition, spatial operators and functions can use a spatial index if defined to avoid sequential full scan of every records.
The current spatial index implementation is built upon lucene-spatial.
The syntax for creating a spatial index on a geometry field is :
CREATE INDEX <name> ON <class-name> (geometry-field) SPATIAL ENGINE LUCENE
Legacy
Before v2.2, OrientDB was able to only index Points. Other Shapes like rectangles and polygons are managed starting from v2.2 (look above). This is the legacy section for databases created before v2.2.
How to create a Spatial Index
The index can be created on a class that has two fields declared as DOUBLE
(latitude
,longitude
) that are the coordinates of the Point.
For example we have a class Place
with 2 double fields latitude
and longitude
. To create the spatial index on Place
use this syntax.
CREATE INDEX Place.l_lon ON Place(latitude,longitude) SPATIAL ENGINE LUCENE
The Index can also be created with the Java Api. Example:
OSchema schema = databaseDocumentTx.getMetadata().getSchema();
OClass oClass = schema.createClass("Place");
oClass.createProperty("latitude", OType.DOUBLE);
oClass.createProperty("longitude", OType.DOUBLE);
oClass.createProperty("name", OType.STRING);
oClass.createIndex("Place.latitude_longitude", "SPATIAL", null, null, "LUCENE", new String[] { "latitude", "longitude" });
How to query the Spatial Index
Two custom operators has been added to query the Spatial Index:
NEAR
: to find all Points near a given location (latitude
,longitude
)WITHIN
: to find all Points that are within a given Shape
NEAR operator
Finds all Points near a given location (latitude
, longitude
).
Syntax
SELECT FROM Class WHERE [<lat-field>,<long-field>] NEAR [lat,lon]
To specify maxDistance
we have to pass a special variable in the context:
SELECT FROM Class WHERE [<lat-field>,<long-field>,$spatial] NEAR [lat,lon,{"maxDistance": distance}]
The maxDistance
field has to be in kilometers, not radians. Results are sorted from nearest to farthest.
To know the exact distance between your Point and the Points matched, use the special variable in the context $distance.
SELECT *, $distance FROM Class WHERE [<lat-field>,<long-field>,$spatial] NEAR [lat,lon,{"maxDistance": distance}]
Examples
Let's take the example we have written before. We have a Spatial Index on Class Place
on properties latitude
and longitude
.
Example: How to find the nearest Place of a given point:
SELECT *,$distance FROM Place WHERE [latitude,longitude,$spatial] NEAR [51.507222,-0.1275,{"maxDistance":1}]
WITHIN operator
Finds all Points that are within a given Shape.
The current release supports only Bounding Box shape |
Syntax
SELECT FROM Class WHERE [<lat field>,<long field>] WITHIN [ [ <lat1>, <lon1> ] , [ <lat2>, <lon2> ] ... ]
Examples
Example with previous configuration:
SELECT * FROM Places WHERE [latitude,longitude] WITHIN [[51.507222,-0.1275],[55.507222,-0.1275]]
This query will return all Places within the given Bounding Box.
Managing Dates
OrientDB treats dates as first class citizens. Internally, it saves dates in the Unix time format. Meaning, it stores dates as a long
variable, which contains the count in milliseconds since the Unix Epoch, (that is, 1 January 1970).
Date and Datetime Formats
In order to make the internal count from the Unix Epoch into something human readable, OrientDB formats the count into date and datetime formats. By default, these formats are:
- Date Format:
yyyy-MM-dd
- Datetime Format:
yyyy-MM-dd HH:mm:ss
In the event that these default formats are not sufficient for the needs of your application, you can customize them through ALTER DATABASE...DATEFORMAT
and DATETIMEFORMAT
commands. For instance,
orientdb> ALTER DATABASE DATEFORMAT "dd MMMM yyyy"
This command updates the current database to use the English format for dates. That is, 14 Febr 2015.
SQL Functions and Methods
To simplify the management of dates, OrientDB SQL automatically parses dates to and from strings and longs. These functions and methods provide you with more control to manage dates:
SQL | Description |
---|---|
DATE() | Function converts dates to and from strings and dates, also uses custom formats. |
SYSDATE() | Function returns the current date. |
.format() | Method returns the date in different formats. |
.asDate() | Method converts any type into a date. |
.asDatetime() | Method converts any type into datetime. |
.asLong() | Method converts any date into long format, (that is, Unix time). |
For example, consider a case where you need to extract only the years for date entries and to arrange them in order. You can use the .format()
method to extract dates into different formats.
orientdb> SELECT @RID, id, date.format('yyyy') AS year FROM Order
--------+----+------+
@RID | id | year |
--------+----+------+
#31:10 | 92 | 2015 |
#31:10 | 44 | 2014 |
#31:10 | 32 | 2014 |
#31:10 | 21 | 2013 |
--------+----+------+
In addition to this, you can also group the results. For instance, extracting the number of orders grouped by year.
orientdb> SELECT date.format('yyyy') AS Year, COUNT(*) AS Total
FROM Order ORDER BY Year
------+--------+
Year | Total |
------+--------+
2015 | 1 |
2014 | 2 |
2013 | 1 |
------+--------+
Dates before 1970
While you may find the default system for managing dates in OrientDB sufficient for your needs, there are some cases where it may not prove so. For instance, consider a database of archaeological finds, a number of which date to periods not only before 1970 but possibly even before the Common Era. You can manage this by defining an era or epoch variable in your dates.
For example, consider an instance where you want to add a record noting the date for the foundation of Rome, which is traditionally referred to as April 21, 753 BC. To enter dates before the Common Era, first run the [ALTER DATABASE DATETIMEFORMAT
] command to add the GG
variable to use in referencing the epoch.
orientdb> ALTER DATABASE DATETIMEFORMAT "yyyy-MM-dd HH:mm:ss GG"
Once you've run this command, you can create a record that references date and datetime by epoch.
orientdb>CREATE VERTEX V SET city = "Rome", date = DATE("0753-04-21 00:00:00 BC")
orientdb>SELECT @RID, city, date FROM V
-------+------+------------------------+ @RID | city | date | -------+------+------------------------+ #9:10 | Rome | 0753-04-21 00:00:00 BC | -------+------+------------------------+
Using .format()
on Insertion
In addition to the above method, instead of changing the date and datetime formats for the database, you can format the results as you insert the date.
orientdb>CREATE VERTEX V SET city = "Rome", date = DATE("yyyy-MM-dd HH:mm:ss GG")
orientdb>SELECT @RID, city, date FROM V
------+------+------------------------+ @RID | city | date | ------+------+------------------------+ #9:4 | Rome | 0753-04-21 00:00:00 BC | ------+------+------------------------+
Here, you again create a vertex for the traditional date of the foundation of Rome. However, instead of altering the database, you format the date field in CREATE VERTEX
command.
Viewing Unix Time
In addition to the formatted date and datetime, you can also view the underlying count from the Unix Epoch, using the asLong()
method for records. For example,
orientdb> SELECT @RID, city, date.asLong() FROM #9:4
------+------+------------------------+
@RID | city | date |
------+------+------------------------+
#9:4 | Rome | -85889120400000 |
------+------+------------------------+
Meaning that, OrientDB represents the date of April 21, 753 BC, as -85889120400000 in Unix time. You can also work with dates directly as longs.
orientdb>CREATE VERTEX V SET city = "Rome", date = DATE(-85889120400000)
orientdb>SELECT @RID, city, date FROM V
-------+------+------------------------+ @RID | city | date | -------+------+------------------------+ #9:11 | Rome | 0753-04-21 00:00:00 BC | -------+------+------------------------+
Use ISO 8601 Dates
According to ISO 8601, Combined date and time in UTC: 2014-12-20T00:00:00. To use this standard change the datetimeformat in the database:
ALTER DATABASE DATETIMEFORMAT "yyyy-MM-dd'T'HH:mm:ss.SSS'Z'"
Transactions
A transaction comprises a unit of work performed within a database management system (or similar system) against a database, and treated in a coherent and reliable way independent of other transactions. Transactions in a database environment have two main purposes:
- to provide reliable units of work that allow correct recovery from failures and keep a database consistent even in cases of system failure, when execution stops (completely or partially) and many operations upon a database remain uncompleted, with unclear status
- to provide isolation between programs accessing a database concurrently. If this isolation is not provided, the program's outcome are possibly erroneous.
A database transaction, by definition, must be atomic, consistent, isolated and durable. Database practitioners often refer to these properties of database transactions using the acronym ACID. --- Wikipedia
OrientDB is an ACID compliant DBMS.
NOTE: OrientDB keeps the transaction on client RAM, so the transaction size is affected by the available RAM (Heap memory) on JVM. For transactions involving many records, consider to split it in multiple transactions.
ACID properties
Atomicity
"Atomicity requires that each transaction is 'all or nothing': if one part of the transaction fails, the entire transaction fails, and the database state is left unchanged. An atomic system must guarantee atomicity in each and every situation, including power failures, errors, and crashes. To the outside world, a committed transaction appears (by its effects on the database) to be indivisible ("atomic"), and an aborted transaction does not happen." - WikiPedia
Consistency
"The consistency property ensures that any transaction will bring the database from one valid state to another. Any data written to the database must be valid according to all defined rules, including but not limited to constraints, cascades, triggers, and any combination thereof. This does not guarantee correctness of the transaction in all ways the application programmer might have wanted (that is the responsibility of application-level code) but merely that any programming errors do not violate any defined rules." - WikiPedia
OrientDB uses the MVCC to assure consistency. The difference between the management of MVCC on transactional and not-transactional cases is that with transactional, the exception rollbacks the entire transaction before to be caught by the application.
Look at this example:
Sequence | Client/Thread 1 | Client/Thread 2 | Version of record X |
---|---|---|---|
1 | Begin of Transaction | ||
2 | read(x) | 10 | |
3 | Begin of Transaction | ||
4 | read(x) | 10 | |
5 | write(x) | 10 | |
6 | commit | 10 -> 11 | |
7 | write(x) | 10 | |
8 | commit | 10 -> 11 = Error, in database x already is at 11 |
Isolation
"The isolation property ensures that the concurrent execution of transactions results in a system state that would be obtained if transactions were executed serially, i.e. one after the other. Providing isolation is the main goal of concurrency control. Depending on concurrency control method, the effects of an incomplete transaction might not even be visible to another transaction." - WikiPedia
OrientDB has different levels of isolation based on settings and configuration:
READ COMMITTED
, the default and the only one available withremote
protocolREPEATABLE READS
, allowed only withplocal
andmemory
protocols. This mode consumes more memory thanREAD COMMITTED
, because any read, query, etc. keep the records in memory to assure the same copy on further access
To change default Isolation Level, use the Java API:
db.begin()
db.getTransaction().setIsolationLevel(OTransaction.ISOLATION_LEVEL.REPEATABLE_READ);
Using remote
access all the commands are executed on the server, so out of transaction scope. Look below for more information.
Look at this examples:
|Sequence| Client/Thread 1 | Client/Thread 2 | |-----|-----|-------|------| |1| Begin of Transaction | | |2| read(x) | | |3| | Begin of Transaction | |4| | read(x) | |5| | write(x) | |6| | commit | |7| read(x) | | |8| commit | |
At operation 7 the client 1 continues to read the same version of x read in operation 2.
|Sequence| Client/Thread 1 | Client/Thread 2 | |-----|-----|-------|------| |1| Begin of Transaction | | |2| read(x) | | |3| | Begin of Transaction | |4| | read(y) | |5| | write(y) | |6| | commit | |7| read(y) | | |8| commit | |
At operation 7 the client 1 reads the version of y which was written at operation 6 by client 2. This is because it never reads y before.
Breaking of ACID properties when using remote protocol and Commands (SQL, Gremlin, JS, etc)
Transactions are client-side only until the commit. This means that if you're using the "remote" protocol the server can't see local changes.
In this scenario you can have different isolation levels with commands. This issue will be solved with OrientDB v3.0 where the transaction will be flushed to the server before to execute the command.
Durability
"Durability means that once a transaction has been committed, it will remain so, even in the event of power loss, crashes, or errors. In a relational database, for instance, once a group of SQL statements execute, the results need to be stored permanently (even if the database crashes immediately thereafter). To defend against power loss, transactions (or their effects) must be recorded in a non-volatile memory." - WikiPedia
Fail-over
An OrientDB instance can fail for several reasons:
- HW problems, such as loss of power or disk error
- SW problems, such as a Operating System crash
- Application problem, such as a bug that crashes your application that is connected to the Orient engine.
You can use the OrientDB engine directly in the same process of your application. This gives superior performance due to the lack of inter-process communication. In this case, should your application crash (for any reason), the OrientDB Engine also crashes.
If you're using an OrientDB Server connected remotely, if your application crashes the engine continue to work, but any pending transaction owned by the client will be rolled back.
Auto-recovery
At start-up the OrientDB Engine checks to if it is restarting from a crash. In this case, the auto-recovery phase starts which rolls back all pending transactions.
OrientDB has different levels of durability based on storage type, configuration and settings.
Transaction types
No Transaction
Default mode. Each operation is executed instantly.
Calls to begin()
, commit()
and rollback()
have no effect.
Optimistic Transaction
This mode uses the well known Multi Version Control System (MVCC) by allowing multiple reads and writes on the same records. The integrity check is made on commit. If the record has been saved by another transaction in the interim, then an OConcurrentModificationException will be thrown. The application can choose either to repeat the transaction or abort it.
NOTE: OrientDB keeps the transaction on client RAM, so the transaction size is affected by the available RAM (Heap) memory on JVM. For transactions involving many records, consider to split it in multiple transactions.
With Graph API transaction begins automatically, with Document API is explicit by using the begin()
method. With Graphs you can change the consistency level.
Example with Document API:
db.open("remote:localhost:7777/petshop");
try{
db.begin(TXTYPE.OPTIMISTIC);
...
// WRITE HERE YOUR TRANSACTION LOGIC
...
db.commit();
}catch( Exception e ){
db.rollback();
} finally{
db.close();
}
In Optimistic transaction new records take temporary Record IDs to avoid to ask to the server a new Record ID every time. Temporary Record IDs have Cluster Id -1 and Cluster Position < -1. When a new transaction begun the counter is reset to -1:-2. So if you create 3 new records you'll have:
- -1:-2
- -1:-3
- -1:-4
At commit time, these temporary records Record IDs will be converted in the final ones.
Pessimistic Transaction
This mode is not yet supported by the engine.
Nested transactions and propagation
OrientDB doesn't support nested transaction. If further begin()
are called after a transaction is already begun, then the current transaction keeps track of call stack to let to the final commit() call to effectively commit the transaction. Look at Transaction Propagation more information.
Record IDs
OrientDB uses temporary Record ID's with transaction as scope that will be transformed to finals once the transactions is successfully committed to the database. This avoid to ask for a free slot every time a client creates a record.
Tuning
In some situations transactions can improve performance, typically in the client/server scenario. If you use an Optimistic Transaction, the OrientDB engine optimizes the network transfer between the client and server, saving both CPU and bandwidth.
For further information look at Transaction tuning to know more.
Distributed environment
Transactions can be committed across a distributed architecture. Look at Distributed Transactions for more information.
Embed the Server
Embedding an OrientDB Server inside a Java application has several advantages and interesting features:
- Java application that runs embedded with the server can bypass the remote connection and use the database directly with local mode. local and remote connections against the same database can work in concurrency: OrientDB will synchronize the access.
- You can use the Console to control it
- You can use the OrientDB Studio
- You can replicate the database across distributed standalone or embedded servers
To embed an OrientDB Server inside a Java application you have to create the OServer
object and use a valid configuration for it.
Requirements
In order to embed the server you need to include the following jar files in the classpath:
orientdb-client-**.jar
orientdb-core-**.jar
orientdb-server-**.jar
orientdb-tools-**.jar
concurrentlinkedhashmap-lru-**.jar
jna-**.jar
lz4-java-**.jar
Starting from version 2.2, please set the MaxDirectMemorySize
parameter. Setting this parameter is required. You can set it to a very high value, e.g. 512g (JVM setting):
-XX:MaxDirectMemorySize=512g
Setting MaxDirectMemorySize
to a very high value should not concern you as it does not mean that OrientDB will consume all 512GB of memory. The size of direct memory consumed by OrientDB is limited by the size of the disk cache (variable storage.diskCache.bufferSize
).
When you start the Server using the provided server.sh
or server.bat
scripts, MaxDirectMemorySize
is set already by those scripts. But when you embed the Server it is required that you set MaxDirectMemorySize
manually.
Note: if you are using a pom file, you may set MaxDirectMemorySize
inside your pom in the following way:
<properties>
<argLine>-XX:MaxDirectMemorySize=512g</argLine>
</properties>
If you start an embedded Server without setting this variable you will get a WARNING message similar to the following:
MaxDirectMemorySize JVM option is not set or has invalid value, that may cause out of memory errors
Include the commands you need
Even if most of the HTTP commands are auto registered assure to have all the commands you need. For example the static content must be registered. This is fundamental if you want to use OrientDB as Web Server providing static content like the Studio app:
<listener protocol="http" port-range="2480-2490" ip-address="0.0.0.0">
<commands>
<command implementation="com.orientechnologies.orient.server.network.protocol.http.command.get.OServerCommandGetStaticContent" pattern="GET|www GET|studio/ GET| GET|*.htm GET|*.html GET|*.xml GET|*.jpeg GET|*.jpg GET|*.png GET|*.gif GET|*.js GET|*.css GET|*.swf GET|*.ico GET|*.txt GET|*.otf GET|*.pjs GET|*.svg">
<parameters>
<entry value="Cache-Control: no-cache, no-store, max-age=0, must-revalidate\r\nPragma: no-cache" name="http.cache:*.htm *.html"/>
<entry value="Cache-Control: max-age=120" name="http.cache:default"/>
</parameters>
</command>
</commands>
</listener>
Use an embedded configuration
import com.orientechnologies.orient.server.OServerMain;
public class OrientDBEmbeddable {
public static void main(String[] args) throws Exception {
OServer server = OServerMain.create();
server.startup(
"<?xml version=\"1.0\" encoding=\"UTF-8\" standalone=\"yes\"?>"
+ "<orient-server>"
+ "<network>"
+ "<protocols>"
+ "<protocol name=\"binary\" implementation=\"com.orientechnologies.orient.server.network.protocol.binary.ONetworkProtocolBinary\"/>"
+ "<protocol name=\"http\" implementation=\"com.orientechnologies.orient.server.network.protocol.http.ONetworkProtocolHttpDb\"/>"
+ "</protocols>"
+ "<listeners>"
+ "<listener ip-address=\"0.0.0.0\" port-range=\"2424-2430\" protocol=\"binary\"/>"
+ "<listener ip-address=\"0.0.0.0\" port-range=\"2480-2490\" protocol=\"http\"/>"
+ "</listeners>"
+ "</network>"
+ "<users>"
+ "<user name=\"root\" password=\"ThisIsA_TEST\" resources=\"*\"/>"
+ "</users>"
+ "<properties>"
+ "<entry name=\"orientdb.www.path\" value=\"C:/work/dev/orientechnologies/orientdb/releases/1.0rc1-SNAPSHOT/www/\"/>"
+ "<entry name=\"orientdb.config.file\" value=\"C:/work/dev/orientechnologies/orientdb/releases/1.0rc1-SNAPSHOT/config/orientdb-server-config.xml\"/>"
+ "<entry name=\"server.cache.staticResources\" value=\"false\"/>"
+ "<entry name=\"log.console.level\" value=\"info\"/>"
+ "<entry name=\"log.file.level\" value=\"fine\"/>"
//The following is required to eliminate an error or warning "Error on resolving property: ORIENTDB_HOME"
+ "<entry name=\"plugin.dynamic\" value=\"false\"/>"
+ "</properties>" + "</orient-server>");
server.activate();
}
}
Once the embedded server is running, clients can connect using the remote connection method. For example in the console, you can connect with:
connect remote:localhost:{port}/{db} {user} {password}
where:
port : the port that the binary server listens on
(first free port from 2424-2430 according to the configuration above)
db : the database name to connect to (defaults to "db" and can be set using <entry name="server.database.path" value="db"/> in the configuration
user : the user to connect with (this is NOT the same as root user in the configuration)
password : the user to connect with (this is NOT the same as root password in the configuration)
Use custom file for configuration
Use a regular File
:
public class OrientDBEmbeddable {
public static void main(String[] args) throws Exception {
OServer server = OServerMain.create();
server.startup(new File("/usr/local/temp/db.config"));
server.activate();
}
}
Use a stream for configuration
Use an InputStream
from the class loader:
public class OrientDBEmbeddable {
public static void main(String[] args) throws Exception {
OServer server = OServerMain.create();
server.startup(getClass().getResourceAsStream("db.config"));
server.activate();
}
}
Use a OServerConfiguration object for configuration
Or an InputStream
from the class loader:
public class OrientDBEmbeddable {
public static void main(String[] args) throws Exception {
OServer server = OServerMain.create();
OServerConfiguration cfg = new OServerConfiguration();
// FILL THE OServerConfiguration OBJECT
server.startup(cfg);
server.activate();
}
}
Shutdown
OrientDB Server creates some threads internally as non-daemon, so they run even if the main application exits. Use the OServer.shutdown()
method to shutdown the server in soft way:
import com.orientechnologies.orient.server.OServerMain;
public class OrientDBEmbeddable {
public static void main(String[] args) throws Exception {
OServer server = OServerMain.create();
server.startup(new File("/usr/local/temp/db.config"));
server.activate();
...
server.shutdown();
}
}
Setting ORIENTDB_HOME
Some functionality will not work properly if the system property 'ORIENTDB_HOME' is not set. You can set it programmatically like this:
import com.orientechnologies.orient.server.OServerMain;
public class OrientDBEmbeddable {
public static void main(String[] args) throws Exception {
String orientdbHome = new File("").getAbsolutePath(); //Set OrientDB home to current directory
System.setProperty("ORIENTDB_HOME", orientdbHome);
OServer server = OServerMain.create();
server.startup(cfg);
server.activate();
}
}
OrientDB Plugins
The OrientDB Server is a customizable platform to build powerful server component and applications.
Since the OrientDB server contains an integrated Web Server what about creating server side applications without the need to have a J2EE and Servlet container? By extending the server you can benefit of the best performance because you don't have many layers but the database and the application reside on the same JVM without the cost of the network and serialization of requests.
Furthermore you can package your application together with the OrientDB server to distribute just a ZIP file containing the entire Application, Web server and Database.
To customize the OrientDB server you have two powerful tools:
To debug the server while you develop new feature follow Debug the server.
Handlers (Server Plugins)
Handlers are plug-ins and starts when OrientDB starts.
To create a new handler create the class and register it in the OrientDB server configuration.
Create the Handler class
A Handler must implements the OServerPlugin interface or extends the OServerPluginAbstract abstract class.
Below an example of a handler that print every 5 seconds a message if the "log" parameters has been configured to be "true":
package orientdb.test;
public class PrinterHandler extends OServerPluginAbstract {
private boolean log = false;
public void config(OServer oServer, OServerParameterConfiguration[] iParams) {
for (OServerParameterConfiguration p : iParams) {
if (p.name.equalsIgnoreCase("log"))
log = true;
}
Orient.getTimer().schedule( new TimerTask() {
public void run() {
if( log )
System.out.println("It's the PrinterHandler!");
}
}, 5000, 5000);
}
public String getName() {
return "PrinterHandler";
}
}
Register the handler
Once created, register it to the server configuration in orientdb-server-config.xml file:
<orient-server>
<handlers>
<handler class="orientdb.test.PrinterHandler">
<parameters>
<parameter name="log" value="true"/>
</parameters>
</handler>
</handlers>
...
Note that you can specify arbitrary parameters in form of name and value. Those parameters can be read by the config() method. In this example a parameter "log" is read. Look upon to the example of handler to know how to read parameters specified in configuration.
Steps to register a function as a Plugin in OrientDB
In this case we'll create a plugin that only registers one function in OrientDB: pow (returns the value of the first argument raised to the power of the second argument). We'll also support Modular exponentiation.
The syntax will be pow(<base>, <power> [, <mod>])
.
- you should have a directory structure like this
.
├─ src
| └─ main
| ├─ assembly
| | └─ assembly.xml
| ├─ java
| | └─ com
| | └─ app
| | └─ OPowPlugin.java
| └─ resources
| └─ plugin.json
|
└─ pom.xml
OPowPlugin.java
package com.app;
import com.orientechnologies.common.log.OLogManager;
import com.orientechnologies.orient.core.command.OCommandContext;
import com.orientechnologies.orient.core.db.record.OIdentifiable;
import com.orientechnologies.orient.core.sql.OSQLEngine;
import com.orientechnologies.orient.core.sql.functions.OSQLFunctionAbstract;
import com.orientechnologies.orient.server.OServer;
import com.orientechnologies.orient.server.config.OServerParameterConfiguration;
import com.orientechnologies.orient.server.plugin.OServerPluginAbstract;
import java.util.ArrayList;
import java.util.List;
public class OPowPlugin extends OServerPluginAbstract {
public OPowPlugin() {
}
public String getName() {
return "pow-plugin";
}
public void startup() {
super.startup();
OSQLEngine.getInstance().registerFunction("pow", new OSQLFunctionAbstract("pow", 2, 3) {
public String getSyntax() {
return "pow(<base>, <power> [, <mod>])";
}
public Object execute(Object iThis, OIdentifiable iCurrentRecord, Object iCurrentResult, final Object[] iParams, OCommandContext iContext) {
if (iParams[0] == null || iParams[1] == null) {
return null;
}
if (!(iParams[0] instanceof Number) || !(iParams[1] instanceof Number)) {
return null;
}
final long base = ((Number) iParams[0]).longValue();
final long power = ((Number) iParams[1]).longValue();
if (iParams.length == 3) { // modular exponentiation
if (iParams[2] == null) {
return null;
}
if (!(iParams[2] instanceof Number)) {
return null;
}
final long mod = ((Number) iParams[2]).longValue();
if (power < 0) {
OLogManager.instance().warn(this, "negative numbers as exponent are not supported");
}
return modPow(base, power, mod);
}
return power > 0 ? pow(base, power) : 1D / pow(base, -power);
}
});
OLogManager.instance().info(this, "pow function registered");
}
private double pow(long base, long power) {
double r = 1;
List<Boolean> bits = bits(power);
for (int i = bits.size() - 1; i >= 0; i--) {
r *= r;
if (bits.get(i)) {
r *= base;
}
}
return r;
}
private double modPow(long base, long power, long mod) {
double r = 1;
List<Boolean> bits = bits(power);
for (int i = bits.size() - 1; i >= 0; i--) {
r = (r * r) % mod;
if (bits.get(i)) {
r = (r * base) % mod;
}
}
return r;
}
private List<Boolean> bits(long n) {
List<Boolean> bits = new ArrayList();
while (n > 0) {
bits.add(n % 2 == 1);
n /= 2;
}
return bits;
}
public void config(OServer oServer, OServerParameterConfiguration[] iParams) {
}
public void shutdown() {
super.shutdown();
}
}
pom.xml
<?xml version="1.0" encoding="UTF-8"?>
<project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/xsd/maven-4.0.0.xsd">
<modelVersion>4.0.0</modelVersion>
<groupId>com.app</groupId>
<artifactId>pow-plugin</artifactId>
<version>2.0.7</version>
<packaging>jar</packaging>
<name>pow-plugin</name>
<properties>
<orientdb.version>2.0.7</orientdb.version>
</properties>
<build>
<plugins>
<plugin>
<artifactId>maven-assembly-plugin</artifactId>
<version>2.4</version>
<configuration>
<descriptors>
<descriptor>src/main/assembly/assembly.xml</descriptor>
</descriptors>
</configuration>
<executions>
<execution>
<id>make-assembly</id>
<!-- this is used for inheritance merges -->
<phase>package</phase>
<!-- bind to the packaging phase -->
<goals>
<goal>single</goal>
</goals>
<configuration></configuration>
</execution>
</executions>
</plugin>
<plugin>
<groupId>org.apache.maven.plugins</groupId>
<artifactId>maven-compiler-plugin</artifactId>
<version>3.1</version>
<configuration>
</configuration>
</plugin>
</plugins>
</build>
<dependencies>
<dependency>
<groupId>com.orientechnologies</groupId>
<artifactId>orientdb-core</artifactId>
<version>${orientdb.version}</version>
<scope>compile</scope>
</dependency>
<dependency>
<groupId>com.orientechnologies</groupId>
<artifactId>orientdb-server</artifactId>
<version>${orientdb.version}</version>
<scope>compile</scope>
</dependency>
</dependencies>
</project>
assembly.xml
<assembly>
<id>dist</id>
<formats>
<format>jar</format>
</formats>
<includeBaseDirectory>false</includeBaseDirectory>
<dependencySets>
<dependencySet>
<outputDirectory/>
<unpack>true</unpack>
<includes>
<include>${groupId}:${artifactId}</include>
</includes>
</dependencySet>
</dependencySets>
</assembly>
plugin.json
{
"name" : "pow-plugin",
"version" : "2.0.7",
"javaClass": "com.app.OPowPlugin",
"parameters" : {},
"description" : "The Pow Plugin",
"copyrights" : "No copyrights"
}
- Build the project and then:
cp target/pow-plugin-2.0.7-dist.jar $ORIENTDB_HOME/plugins/
You should see the following in OrientDB server log:
INFO Installing dynamic plugin 'pow-plugin-2.0.7-dist.jar'... [OServerPluginManager]
INFO pow function registered [OPowPlugin]
And now you can:
orientdb {db=Pow}> select pow(2,10)
----+------+------
# |@CLASS|pow
----+------+------
0 |null |1024.0
----+------+------
orientdb {db=Pow}> select pow(2,10,5)
----+------+----
# |@CLASS|pow
----+------+----
0 |null |4.0
----+------+----
This small project is available here.
Creating a distributed change manager
As more complete example let's create a distributed record manager by installing hooks to all the server's databases and push these changes to the remote client caches.
public class DistributedRecordHook extends OServerHandlerAbstract implements ORecordHook {
private boolean log = false;
public void config(OServer oServer, OServerParameterConfiguration[] iParams) {
for (OServerParameterConfiguration p : iParams) {
if (p.name.equalsIgnoreCase("log"))
log = true;
}
}
public void onAfterClientRequest(final OClientConnection iConnection, final byte iRequestType) {
if (iRequestType == OChannelBinaryProtocol.REQUEST_DB_OPEN)
iConnection.database.registerHook(this);
else if (iRequestType == OChannelBinaryProtocol.REQUEST_DB_CLOSE)
iConnection.database.unregisterHook(this);
}
public boolean onTrigger(TYPE iType, ORecord<?> iRecord) {
try {
if (log)
System.out.println("Broadcasting record: " + iRecord + "...");
OClientConnectionManager.instance().broadcastRecord2Clients((ORecordInternal<?>) iRecord, null);
} catch (Exception e) {
e.printStackTrace();
}
return false;
}
public String getName() {
return "DistributedRecordHook";
}
}
Custom commands
Custom commands are useful when you want to add behavior or business logic at the server side.
A Server command is a class that implements the OServerCommand interface or extends one of the following abstract classes:
- OServerCommandAuthenticatedDbAbstract if the command requires an authentication at the database
- OServerCommandAuthenticatedServerAbstract if the command requires an authentication at the server
The Hello World Web
To learn how to create a custom command, let's begin with a command that just returns "Hello world!".
OrientDB follows the convention that the command name is:
OServerCommand<method><name>
Where:
- method is the HTTP method and can be: GET, POST, PUT, DELETE
- name is the command name
In our case the class name will be "OServerCommandGetHello". We want that the use must be authenticated against the database to execute it as any user.
Furthermore we'd like to receive via configuration if we must display the text in Italic or not, so for this purpose we'll declare a parameter named "italic" of type boolean (true or false).
package org.example;
public class OServerCommandGetHello extends OServerCommandAuthenticatedDbAbstract {
// DECLARE THE PARAMETERS
private boolean italic = false;
public OServerCommandGetHello(final OServerCommandConfiguration iConfiguration) {
// PARSE PARAMETERS ON STARTUP
for (OServerEntryConfiguration par : iConfiguration.parameters) {
if (par.name.equals("italic")) {
italic = Boolean.parseBoolean(par.value);
}
}
}
public boolean execute(final OHttpRequest iRequest, OHttpResponse iResponse) throws Exception {
// CHECK THE SYNTAX. 3 IS THE NUMBER OF MANDATORY PARAMETERS
String[] urlParts = checkSyntax(iRequest.url, 3, "Syntax error: hello/<database>/<name>");
// TELLS TO THE SERVER WHAT I'M DOING (IT'S FOR THE PROFILER)
iRequest.data.commandInfo = "Salutation";
iRequest.data.commandDetail = "This is just a test";
// GET THE PARAMETERS
String name = urlParts[2];
// CREATE THE RESULT
String result = "Hello " + name;
if (italic) {
result = "<i>" + result + "</i>";
}
// SEND BACK THE RESPONSE AS TEXT
iResponse.send(OHttpUtils.STATUS_OK_CODE, "OK", null, result, OHttpUtils.CONTENT_TEXT_PLAIN);
// RETURN ALWAYS FALSE, UNLESS YOU WANT TO EXECUTE COMMANDS IN CHAIN
return false;
}
public String[] getNames() {
return new String[]{"GET|hello/*","POST|hello/*"};
}
}
Once created the command you need to register them through the orientdb-server-config.xml file. Put a new tag <command>
under the tag commands
of <listener>
with attribute protocol="http"
:
...
<listener protocol="http" port-range="2480-2490" ip-address="0.0.0.0">
<commands>
<command implementation="org.example.OServerCommandGetHello" pattern="GET|hello/*">
<parameters>
<entry name="italic" value="true"/>
</parameters>
</command>
</commands>
</listener>
Where:
- implementation is the full class name of the command
- pattern is how the command is called in the format:
<HTTP-method>|<name>
. In this case it's executed on HTTP GET with the URL:/<name>
- parameters specify parameters to pass to the command on startup
- entry is the parameter pair name/value
To test it open a browser at this address:
http://localhost:2480/hello/<your-db-name>/Luca
You will see:
Hello Luca
Complete example
Below a more complex example taken by official distribution. It is the command that executes queries via HTTP. Note how to get a database instance to execute operation against the database:
public class OServerCommandGetQuery extends OServerCommandAuthenticatedDbAbstract {
private static final String[] NAMES = { "GET|query/*" };
public boolean execute(OHttpRequest iRequest, OHttpResponse iResponse) throws Exception {
String[] urlParts = checkSyntax(
iRequest.url,
4,
"Syntax error: query/<database>/sql/<query-text>[/<limit>][/<fetchPlan>].<br/>Limit is optional and is setted to 20 by default. Set expressly to 0 to have no limits.");
int limit = urlParts.length > 4 ? Integer.parseInt(urlParts[4]) : 20;
String fetchPlan = urlParts.length > 5 ? urlParts[5] : null;
String text = urlParts[3];
iRequest.data.commandInfo = "Query";
iRequest.data.commandDetail = text;
ODatabaseDocumentTx db = null;
List<OIdentifiable> response;
try {
db = getProfiledDatabaseInstance(iRequest);
response = (List<OIdentifiable>) db.command(new OSQLSynchQuery<OIdentifiable>(text, limit).setFetchPlan(fetchPlan)).execute();
} finally {
if (db != null) {
db.close();
}
}
iResponse.writeRecords(response, fetchPlan);
return false;
}
public String[] getNames() {
return NAMES;
}
}
Include JARS in the classpath
If your extensions need additional libraries put the additional jar files under the /lib
folder of the server installation.
Debug the server
To debug your plugin you can start your server in debug mode.
Parameter | Value |
---|---|
Main class | com.orientechnologies.orient.server.OServerMain |
JVM parameters | -server -DORIENTDB_HOME=/opt/orientdb -Dorientdb.www.path=src/site -Djava.util.logging.config.file=${ORIENTDB_HOME}/config/orientdb-server-log.properties -Dorientdb.config.file=${ORIENTDB_HOME}/config/orientdb-server-config.xml |
Automatic Backup Server Plugin
Using this server plugin, OrientDB executes regular backups on the databases. It implements the Java class:
com.orientechnologies.orient.server.handler.OAutomaticBackup
Plugin Configuration
Beginning with version 2.2, OrientDB manages the server plugin configuration from a separate JSON. You can update this file manually or through OrientDB Studio.
To enable automatic backups, use the following <handler>
section in the config/orientdb-server-config.xml
configuration file:
<!-- AUTOMATIC BACKUP, TO TURN ON SET THE 'ENABLED' PARAMETER TO 'true' -->
<handler class="com.orientechnologies.orient.server.handler.OAutomaticBackup">
<parameters>
<parameter name="enabled" value="false"/>
<!-- LOCATION OF JSON CONFIGURATION FILE -->
<parameter name="config" value="${ORIENTDB_HOME}/config/automatic-backup.json"/>
</parameters>
</handler>
This section tells the OrientDB server to read the file at $ORIENTDB_HOME/config/automatic-backup.json
for the automatic backup configuration.
{
"enabled": true,
"mode": "FULL_BACKUP",
"exportOptions": "",
"delay": "4h",
"firstTime": "23:00:00",
"targetDirectory": "backup",
"targetFileName": "${DBNAME}-${DATE:yyyyMMddHHmmss}.zip",
"compressionLevel": 9,
"bufferSize": 1048576
}
"enabled"
Defines whether it uses automatic backups. The supported values are:true
Enables automatic backups.false
Disables automatic backups. This is the default setting.
"mode"
Defines the backup mode. The supported values are:"FULL_BACKUP"
Executes a full backup. Prior to version 2.2, this was the only mode available. This operation blocks the database."INCREMENTAL_BACKUP"
Executes an incremental backup. This is available only in the Enterprise Edition. It uses one directory per database. This operation doesn't block the database."EXPORT"
Executes an database export, using gziped JSON format. This operation is not blocking.
"exportOptions"
Defines export options to use with that mode. This feature was introduced in version 2.2."delay"
Defines the delay time for each backup. Supports the following suffixes:ms
Delay measured in milliseconds.s
Delay measured in seconds.m
Delay measured in minutes.h
Delay measured in hours.d
Delay measured in days.
"firstTime"
Defines when to initiate the first backup in the schedule. It uses the format ofHH:mm:ss
in the GMT time zone, on the current day."targetDirectory"
Defines the target directory to write backups. By default, it is set to thebackup/
directory."targetFileName"
Defines the target filename. This parameter supports the use of the following variables, (that is,"${DBNAME}-backup.zip"
producesmydatabase-backup.zip
):${DBNAME}
Renders the database name.${DATE}
Renders the current date, using the Java DateTime syntax format.
"dbInclude"
Defines in a list the databases to include in the automatic backups. If empty, it backs up all databases."dbExclude"
Defines in a list the databases to exclude from the automatic backups."bufferSize"
Defines the in-memory buffer sizes to use in compression. By default, it is set to1MB
. Larger buffers mean faster backups, but they in turn consume more RAM."compressionLevel"
Defines the compression level for the resulting ZIP file. By default it is set to the maximum level of9
. Set it to a lower value if you find that the backup takes too much time.
Legacy Plugin Configuration
In versions prior to 2.2, the only option in configuring automatic backups is to use the config/orientdb-server-config.xml
configuration file. Beginning with version 2.2 you can manage automatic backup configuration through a separate JSON file or use the legacy approach.
The example below configures automatic backups/exports on the database as a Server Plugin.
<!-- AUTOMATIC BACKUP, TO TURN ON SET THE 'ENABLED' PARAMETER TO 'true' -->
<handler class="com.orientechnologies.orient.server.handler.OAutomaticBackup">
<parameters>
<parameter name="enabled" value="false" />
<!-- CAN BE: FULL_BACKUP, INCREMENTAL_BACKUP, EXPORT -->
<parameter name="mode" value="FULL_BACKUP"/>
<!-- OPTION FOR EXPORT -->
<parameter name="exportOptions" value=""/>
<parameter name="delay" value="4h" />
<parameter name="target.directory" value="backup" />
<!-- ${DBNAME} AND ${DATE:} VARIABLES ARE SUPPORTED -->
<parameter name="target.fileName" value="${DBNAME}-${DATE:yyyyMMddHHmmss}.zip" />
<!-- DEFAULT: NO ONE, THAT MEANS ALL DATABASES.
USE COMMA TO SEPARATE MULTIPLE DATABASE NAMES -->
<parameter name="db.include" value="" />
<!-- DEFAULT: NO ONE, THAT MEANS ALL DATABASES.
USE COMMA TO SEPARATE MULTIPLE DATABASE NAMES -->
<parameter name="db.exclude" value="" />
<parameter name="compressionLevel" value="9"/>
<parameter name="bufferSize" value="1048576"/>
</parameters>
</handler>
enabled
Defines whether it uses automatic backups. Supported values are:true
Enables automatic backups.false
Disables automatic backups. This is the default setting.
mode/>
Defines the backup mode. Supported values are:FULL_BACKUP
Executes a full backup. For versions prior to 2.2, this is the only option available. This operation blocks the database.INCREMENTAL_BACKUP
Executes an incremental backup. Uses one directory per database. This operation doesn't block the database.- *
EXPORT
Executes an export of the database in gzipped JSON format, instead of a backup. This operation doesn't block the database.
exportOptions
Defines export options to use with that mode. This feature was introduced in version 2.2.delay
Defines the delay time. Supports the following suffixes:ms
Delay measured in milliseconds.s
Delay measured in seconds.m
Delay measured in minutes.h
Delay measured in hours.d
Delay measured in days.
firstTime
Defines when to initiate the first backup in the schedule. It uses the format ofHH:mm:ss
in the GMT time zone, on the current day.target.directory
Defines the target directory to write backups. By default, it is set to thebackup/
directory.target.fileName
Defines the target file name. The parameter supports the use of the following variables, (that is,<parameter name="target.filename" value="${DBNAME}-backup.zip"/>
produces amydatabase-backup.zip
file).${DBNAME}
Renders the database name.${DATE}
Renders the current date, using the Java DateTime syntax format.
db.include
Defines in a list the databases to include in the automatic backups. If left empty, it backs up all databases.db.exclude
Defines in a list the databases to exclude from automatic backups.bufferSize
Defines the in-memory buffer sizes to use in compression. By default it is set to1MB
. Larger buffers mean faster backups, but use more RAM. This feature was introduced in version 1.7.compressionLevel
Defines the compression level for the resulting ZIP file. By default, it is set to the maximum level of9
. Set it to a lower value if you find that the backup takes too much time.
SysLog Plugin
Java class implementation:
com.orientechnologies.security.syslog.ODefaultSyslog
Available since: v. 2.2.0.
Introduction
Allows sending event logs to the Operating System's SYSLOG daemon.
Configuration
This plugin is configured as a Server plugin. The plugin can be easily configured by changing parameters in the orientdb-server-config.xml
file.:
Name | Description | Type | Example | Since |
---|---|---|---|---|
enabled | true to turn on, false (default) is turned off | boolean | true | 2.2.0 |
debug | Enables debug mode | boolean | false | 2.2.0 |
hostname | The hostname of the syslog daemon | string | localhost | 2.2.0 |
port | The UDP port of the syslog daemon | integer | 514 | 2.2.0 |
appName | The name of the application submitting the events to SysLog | string | OrientDB | 2.2.0 |
Default configuration in orientdb-server-config.xml
. Example:
<!-- SYS LOG CONNECTOR, TO TURN ON SET THE 'ENABLED' PARAMETER TO 'true' -->
<handler class="com.orientechnologies.security.syslog.ODefaultSyslog">
<parameters>
<parameter name="enabled" value="true"/>
<parameter name="debug" value="false"/>
<parameter name="hostname" value="localhost"/>
<parameter name="port" value="514"/>
<parameter name="appName" value="OrientDB"/>
</parameters>
</handler>
Usage
Look at Security Config.
Mail Plugin
Java class implementation:
com.orientechnologies.orient.server.plugin.mail.OMailPlugin
Available since: v. 1.2.0.
Introduction
Allows to send (and in future read) emails.
Configuration
This plugin is configured as a Server handler. The plugin can be configured in easy way by changing parameters:
Name | Description | Type | Example | Since |
---|---|---|---|---|
enabled | true to turn on, false (default) is turned off | boolean | true | 1.2.0 |
profile.<name>.mail.smtp.host | The SMTP host name or ip-address | string | smtp.gmail.com | 1.2.0 |
profile.<name>.mail.smtp.port | The SMTP port | number | 587 | 1.2.0 |
profile.<name>.mail.smtp.auth | Authenticate in SMTP | boolean | true | 1.2.0 |
profile.<name>.mail.smtp.starttls.enable | Enable the starttls | boolean | true | 1.2.0 |
profile.<name>.mail.smtp.user | The SMTP username | string | yoda@starwars.com | 1.2.0 |
profile.<name>.mail.from | The source's email address | string | yoda@starwars.com | 1.7 |
profile.<name>.mail.smtp.password | The SMTP password | string | UseTh3F0rc3 | 1.2.0 |
profile.<name>.mail.date.format | The date format to use, default is "yyyy-MM-dd HH:mm:ss" | string | yyyy-MM-dd HH:mm:ss | 1.2.0 |
Default configuration in orientdb-server-config.xml. Example:
<!-- MAIL, TO TURN ON SET THE 'ENABLED' PARAMETER TO 'true' -->
<handler
class="com.orientechnologies.orient.server.plugin.mail.OMailPlugin">
<parameters>
<parameter name="enabled" value="true" />
<!-- CREATE MULTIPLE PROFILES WITH profile.<name>... -->
<parameter name="profile.default.mail.smtp.host" value="smtp.gmail.com"/>
<parameter name="profile.default.mail.smtp.port" value="587" />
<parameter name="profile.default.mail.smtp.auth" value="true" />
<parameter name="profile.default.mail.smtp.starttls.enable" value="true" />
<parameter name="profile.default.mail.from" value="test@gmail.com" />
<parameter name="profile.default.mail.smtp.user" value="test@gmail.com" />
<parameter name="profile.default.mail.smtp.password" value="mypassword" />
<parameter name="profile.default.mail.date.format" value="yyyy-MM-dd HH:mm:ss" />
</parameters>
</handler>
Usage
The message is managed as a map of properties containing all the fields those are part of the message.
Supported message properties:
Name | Description | Mandatory | Example | Since |
---|---|---|---|---|
from | source email address | No | to : "first@mail.com", "second@mail.com" | 1.7 |
to | destination addresses separated by commas | Yes | to : "first@mail.com", "second@mail.com" | 1.2.0 |
cc | Carbon copy addresses separated by commas | No | cc: "first@mail.com", "second@mail.com" | 1.2.0 |
bcc | Blind Carbon Copy addresses separated by commas | No | bcc : "first@mail.com", "second@mail.com" | 1.2.0 |
subject | The subject of the message | No | subject : "This Email plugin rocks!" | 1.2.0 |
message | The message's content | Yes | message : "Hi, how are you mate?" | 1.2.0 |
date | The subject of the message. Pass a java.util.Date object or a string formatted following the rules specified in "mail.date.format" configuration parameter or "yyyy-MM-dd HH:mm:ss" is taken | No, if not specified current date is assumed | date : "2012-09-25 13:20:00" | 1.2.0 |
attachments | The files to attach | No | 1.2.0 |
From Server-Side Functions
The Email plugin install a new variable in the server-side function's context: "mail". "profile" attribute is the profile name in configuration.
Example to send an email writing a function in JS:
mail.send({
profile : "default",
to: "orientdb@ruletheworld.com",
cc: "yoda@starwars.com",
bcc: "darthvader@starwars.com",
subject: "The EMail plugin works",
message : "Sending email from OrientDB Server is so powerful to build real web applications!"
});
On Nashorn (>= Java8) the mapping of JSON to Map is not implicit. Use this:
mail.send( new java.util.HashMap{
profile : "default",
to: "orientdb@ruletheworld.com",
cc: "yoda@starwars.com",
bcc: "darthvader@starwars.com",
subject: "The EMail plugin works",
message : "Sending email from OrientDB Server is so powerful to build real web applications!"
});
From Java
OMailPlugin plugin = OServerMain.server().getPlugin("mail");
Map<String, Object> message = new HashMap<String, Object>();
message.put("profile", "default");
message.put("to", "orientdb@ruletheworld.com");
message.put("cc", "yoda@starts.com,yoda-beach@starts.com");
message.put("bcc", "darthvader@starwars.com");
message.put("subject", "The EMail plugin works");
message.put("message", "Sending email from OrientDB Server is so powerful to build real web applications!");
plugin.send(message);
JMX plugin
Java class implementation:
com.orientechnologies.orient.server.handler.OJMXPlugin
Available since: v. 1.2.0.
Introduction
Expose the OrientDB server configuration through JMX protocol. This task is configured as a Server handler. The task can be configured in easy way by changing parameters:
- enabled: true to turn on, false (default) is turned off
- profilerManaged: manage the Profiler instance
Default configuration in orientdb-server-config.xml
<!-- JMX SERVER, TO TURN ON SET THE 'ENABLED' PARAMETER TO 'true' -->
<handler class="com.orientechnologies.orient.server.handler.OJMXPlugin">
<parameters>
<parameter name="enabled" value="false" />
<parameter name="profilerManaged" value="true" />
</parameters>
</handler>
Rexster
Rexster provides a RESTful shell to any Blueprints-complaint graph database. This HTTP web service provides: a set of standard low-level GET, POST, and DELETE methods, a flexible extension model which allows plug-in like development for external services (such as ad-hoc graph queries through Gremlin), and a browser-based interface called The Dog House.
A graph database hosted in the OrientDB can be configured in Rexster and then accessed using the standard RESTful interface powered by the Rexster web server.
Installation
You can get the latest stable release of Rexster from its Download Page. The latest stable release when this page was last updated was 2.5.0.
Or you can build a snapshot by executing the following Git and Maven commands:
git clone https://github.com/tinkerpop/rexster.git
cd rexster
mvn clean install
Rexster is distributed as a zip file (also the building process creates a zip file) hence the installation consist of unzipping the archive in a directory of your choice. In the following sections, this directory is referred to as $REXSTER_HOME.
After unzipping the archive, you should copy orient-client.jar and orient-enterprise.jar in $REXSTER_HOME/ext. Make sure you use the same version of OrientDB as those used by Rexster. For example Rexster 2.5.0 uses OrientDB 1.7.6.
You can find more details about Rexster installation at the Getting Started page.
Configuration
Refer to Rexster's Configuration page and OrientDB specific configuration page for the latest details.
Synopsis
The Rexster configuration file rexster.xml is used to configure parameters such as: TCP ports used by Rexster server modules to listen for incoming connections; character set supported by the Rexster REST requests and responses; connection parameters of graph instances.
In order to configure Rexster to connect to your OrientDB graph, locate the rexster.xml in the Rexster directory and add the following snippet of code:
<rexster>
...
<graphs>
...
<graph>
<graph-enabled>true</graph-enabled>
<graph-name>my-orient-graph</graph-name>
<graph-type>orientgraph</graph-type>
<graph-file>url-to-your-db</graph-file>
<properties>
<username>user</username>
<password>pwd</password>
</properties>
</graph>
...
</graphs>
</rexster>
In the configuration file, there could be a sample graph
element for an OrientDB instance (<graph-name>orientdbsample<graph-name>
): you might edit it according to your needs.
The <graph-name>
element must be unique within the list of configured graphs and reports the name used to identify your graph.
The <graph-enabled>
element states whether the graph should be loaded and managed by Rexster. Setting its contents to false
will prevent that graph from loading to Rexster; setting explicitly to true
the graph will be loaded.
The <graph-type>
element reports the type of graph by using an identifier (orientgraph
for an OrientDB Graph instance) or the full name of the class that implements the GraphConfiguration interface
(com.tinkerpop.rexster.OrientGraphConfiguration for an OrientDB Graph).
The <graph-file>
element reports the URL to the OrientDB database Rexster is expected to connect to:
plocal:*path-to-db*
, if the graph can be accessed over the file system (e.g.plocal:/tmp/graph/db
)remote:*url-to-db*
, if the graph can be accessed over the network and/or if you want to enable multiple accesses to the graph (e.g.remote:localhost/mydb
)memory:*db-name*
, if the graph resides in memory only. Updates to this kind of graph are never persistent and when the OrientDB server ends the graph is lost
The <username>
and <password>
elements reports the credentials to access the graph (e.g. admin
admin
).
Run
Note: only Rexster 0.5-SNAPSHOT and further releases work with OrientDB GraphEd
In this section we present a step-by-step guide to Rexster-ify an OrientDB graph.
We assume that:
- you created a Blueprints enabled graph called orientGraph using the class
com.tinkerpop.blueprints.pgm.impls.orientdb.OrientGraph
- you inserted in the Rexster configuration file a
<graph>
element with the<graph-name>
element set tomy-orient-graph
and thegraph-file
element set toremote:orienthost/orientGraph
(if you do not remember how to do this, go back to the Configuration section).
- Be sure that the OrientDB server is running and you have properly configured the
<graph-file>
location and the access credentials of your graph. - Execute the startup script ($REXSTER_HOME/bin/rexster.bat or $REXSTER_HOME/bin/rexster.sh)
- The shell console appears and you should see the following log message (line 10 states that the OrientDB graph instance has been loaded):
[INFO] WebServer - .:Welcome to Rexster:.
[INFO] GraphConfigurationContainer - Graph emptygraph - tinkergraph[vertices:0 edges:0] loaded
[INFO] RexsterApplicationGraph - Graph [tinkergraph] - configured with allowable namespace [tp:gremlin]
[INFO] GraphConfigurationContainer - Graph tinkergraph - tinkergraph[vertices:6 edges:6] loaded
[INFO] RexsterApplicationGraph - Graph [tinkergraph-readonly] - configured with allowable namespace [tp:gremlin]
[INFO] GraphConfigurationContainer - Graph tinkergraph-readonly - (readonly)tinkergraph[vertices:6 edges:6] loaded
[INFO] RexsterApplicationGraph - Graph [gratefulgraph] - configured with allowable namespace [tp:gremlin]
[INFO] GraphConfigurationContainer - Graph gratefulgraph - tinkergraph[vertices:809 edges:8049] loaded
[INFO] GraphConfigurationContainer - Graph sailgraph - sailgraph[memorystore] loaded
[INFO] GraphConfigurationContainer - Graph my-orient-graph - orientgraph[remote:orienthost/orientGraph] loaded
[INFO] GraphConfigurationContainer - Graph neo4jsample - not enabled and not loaded.
[INFO] GraphConfigurationContainer - Graph dexsample - not enabled and not loaded.
[INFO] MapResultObjectCache - Cache constructed with a maximum size of 1000
[INFO] WebServer - Web Server configured with com..sun..jersey..config..property..packages: com.tinkerpop.rexster
[INFO] WebServer - No servlet initialization parameters passed for configuration: admin-server-configuration
[INFO] WebServer - Rexster Server running on: [http://localhost:8182]
[INFO] WebServer - Dog House Server running on: [http://localhost:8183]
[INFO] ShutdownManager$ShutdownSocketListener - Bound shutdown socket to /127.0.0.1:8184. Starting listener thread for shutdown requests.
- Now you can use Rexster REST API and The Dog House web application to retrieve and modify the data stored in the OrientDB graph.
Gephi Visual Tool
Introduction
Gephi is a visual tool to manipulate and analyze graphs. http://gephi.org is an Open Source project. Take a look at the amazing features.
Gephi can be used to analyze graphs extracted from OrientDB. There are 2 level of integration:
- the Streaming plugin that calls OrientDB server via HTTP. OrientDB exposes the new "/gephi" command in HTTP GET method that executes a query and returns the result set in "gephi" format.
- Gephi importer for Blueprints
In this mini guide we will take a look at the first one: the streaming plugin.
For more information:
Getting started
Before to start assure you've OrientDB 1.1.0-SNAPSHOT or greater.
Download and install
- To download Gephi goto: http://gephi.org/users/download/
- Install it, depends on your OS
- Run Gephi
- Click on the menu Tools -> Plugins
- Click on the tab Available Plugins
- Select the plugin Graph Streaming, click on the Install button and wait the plugin is installed
Import a graph in Gephi
Before to import a graph assure a OrientDB server instance is running somewhere. For more information watch this video.
- Go to the Overview view (click on Overview top left button)
- Click on the Streaming tab on the left
- Click on the big + green button
- Insert as Source URL the query you want to execute. Example:
http://localhost:2480/gephi/demo/sql/select%20from%20v/100
(below more information about the syntax of query) - Select as Stream type the JSON format (OrientDB talks in JSON)
- Enable the Use Basic Authentication and insert the user and password of OrientDB database you want to access. The default user is "admin" as user and password
- Click on OK button
Executing a query
The OrientDB's "/gephi" HTTP command allow to execute any query. The format is:
http://<host>:<port>/gephi/<database>/<language>/<query>[/<limit>]
Where:
host
is the host name or the ip address where the OrientDB server is running. If you're executing OrientDB on the same machine where Gephi is running use "localhost"port
is the port number where the OrientDB server is running. By default is 2480.database
is the database namelanguage
query
, the query text following the URL encoding rules. For example to use the spaces use%20
, so the queryselect from v
becomesselect%20from%20v
limit
, optional, set the limit of the result set. If not defined 20 is taken by default.-1
means no limits
SQL Graph language
To use the OrientDB's SQL language use sql
as language. For more information look at the SQL-Syntax.
For example, to return the first 1,000 vertices (class V) with outgoing connections the query would be:
SELECT FROM V WHERE out.size() > 0
Executed on "localhost" against the "demo" database + encoding becomes:
http://localhost:2480/gephi/demo/sql/select%20from%20V%20where%20out.size()%20%3E%200/1000
GREMLIN language
To use the powerful GREMLIN language to retrieve the graph or a portion of it use gremlin
as language. For more information look at the GREMLIN syntax.
For example, to return the first 100 vertices:
g.V[0..99]
Executed on "localhost" against the "demo" database + encoding becomes:
http://localhost:2480/gephi/demo/gremlin/g.V%5B0..99%5D/-1
spider-box
spider-box is not really a "plug-in", but more a quick way to set up an environment to play with OrientDB in a local VM. It requires a virtualization system like Virtualbox, VMWare Fusion or Parallels and the provisioning software Vagrant.
Once installed, you can very quickly start playing with the newest version of OrientDB Studio or the console. Or even start developing software with OrientDB as the database.
spider-box is configured mainly to build a PHP development environment. But, since it is built on Puphpet, you can easily change the configuration, so Python or even node.js is also installed. Ruby is installed automatically.
If you have questions about changing configuration or using spider-box, please do ask in an issue in the spider-box repo.
Have fun playing with OrientDB and spider-box!
Note: Neo4j and Gremlin Server are also installed, when you vagrant up
spider-box.
Plugins
If you're looking for drivers or JDBC connector go to Programming-Language-Bindings.
- OrientDB Spring Data is the official Spring Data Plugin for both Graph and Document APIs
- spring-orientdb is an attempt to provide a PlatformTransactionManager for OrientDB usable with the Spring Framework, in particular with @Transactional annotation. Apache 2 license
- Spring Session OrientDB is a Spring Session extension for OrientDB.
- Play Framework 2.1 PLAY-WITH-ORIENTDB plugin
- Play Framework 2.1 ORIGAMI plugin
- Play Framework 1.x ORIENTDB plugin
- Frames-OrientDB Plugin Play Framework 2.x Frames-OrientDB plugin is a Java O/G mapper for the OrientDB with the Play! framework 2. It is used with the TinkerPop Frames for O/G mapping.
With proper mark-up/logic separation, a POJO data model, and a refreshing lack of XML, Apache Wicket makes developing web-apps simple and enjoyable again. Swap the boilerplate, complex debugging and brittle code for powerful, reusable components written with plain Java and HTML.
Guice (pronounced 'juice') is a lightweight dependency injection framework for Java 6 and above, brought to you by Google. OrientDB Guice plugin allows to integrate OrientDB inside Guice. Features:
- Integration through guice-persist (UnitOfWork, PersistService, @Transactional, dynamic finders supported)
- Support for document, object and graph databases
- Database types support according to classpath (object and graph db support activated by adding jars to classpath)
- Auto mapping entities in package to db scheme or using classpath scanning to map annotated entities
- Auto db creation
- Hooks for schema migration and data initialization extensions
- All three database types may be used in single unit of work (but each type will use its own transaction)
Vert.x is a lightweight, high performance application platform for the JVM that's designed for modern mobile, web, and enterprise applications. Vert.x Persistor Module for Tinkerpop-compatible Graph Databases like OrientDB.
Gephi Visual tool usage with OrientDB and the Blueprints importer
OrientDB session store for Connect
[Puppet module](https://github.com/example42/puppet-orientdb)
Chef
Apache Tomcat realm plugin by Jonathan Tellier
Shibboleth connector by Jonathan Tellier. The Shibboleth System is a standards based, open source software package for web single sign-on across or within organizational boundaries. It allows sites to make informed authorization decisions for individual access of protected online resources in a privacy-preserving manner
Griffon plugin, Apache 2 license
JCA connectors
- OPS4J Orient provides a JCA resource adapter for integrating OrientDB with Java EE 6 servers
- OrientDB JCA connector to access to OrientDB database via JCA API + XA Transactions
Pacer plugin by Paul Dlug. Pacer is a JRuby graph traversal framework built on the Tinkerpop stack. This plugin enables full OrientDB graph support in Pacer.
EventStore for Axonframework, which uses fully transactional (full ACID support) NoSQL database OrientDB. Axon Framework helps build scalable, extensible and maintainable applications by supporting developers apply the Command Query Responsibility Segregation (CQRS) architectural pattern
Accessing OrientDB using Slick
Jackrabbit module to use OrientDB as backend.
orientqb is a builder for OSQL query language written in Java. orientqb has been thought to help developers in writing complex queries dynamically and aims to be simple but powerful.
Multi-Model
The OrientDB engine supports Graph, Document, Key/Value, and Object models, so you can use OrientDB as a replacement for a product in any of these categories. However, the main reason why users choose OrientDB is because of its true Multi-Model DBMS abilities, which combine all the features of the four models into the core. These abilities are not just interfaces to the database engine, but rather the engine itself was built to support all four models. This is also the main difference to other multi-model DBMSs, as they implement an additional layer with an API, which mimics additional models. However, under the hood, they're truly only one model, therefore they are limited in speed and scalability.
The Document Model
The data in this model is stored inside documents. A document is a set of key/value pairs (also referred to as fields or properties), where the key allows access to its value. Values can hold primitive data types, embedded documents, or arrays of other values. Documents are not typically forced to have a schema, which can be advantageous, because they remain flexible and easy to modify. Documents are stored in collections, enabling developers to group data as they decide. OrientDB uses the concepts of "classes" and "clusters" as its form of "collections" for grouping documents. This provides several benefits, which we will discuss in further sections of the documentation.
OrientDB's Document model also adds the concept of a "LINK" as a relationship between documents. With OrientDB, you can decide whether to embed documents or link to them directly. When you fetch a document, all the links are automatically resolved by OrientDB. This is a major difference to other Document Databases, like MongoDB or CouchDB, where the developer must handle any and all relationships between the documents herself.
The table below illustrates the comparison between the relational model, the document model, and the OrientDB document model:
Relational Model | Document Model | OrientDB Document Model |
---|---|---|
Table | Collection | Class or Cluster |
Row | Document | Document |
Column | Key/value pair | Document field |
Relationship | not available | Link |
The Graph Model
A graph represents a network-like structure consisting of Vertices (also known as Nodes) interconnected by Edges (also known as Arcs). OrientDB's graph model is represented by the concept of a property graph, which defines the following:
-
Vertex - an entity that can be linked with other Vertices and has the following mandatory properties:
- unique identifier
- set of incoming Edges
- set of outgoing Edges
-
Edge - an entity that links two Vertices and has the following mandatory properties:
- unique identifier
- link to an incoming Vertex (also known as head)
- link to an outgoing Vertex (also known as tail)
- label that defines the type of connection/relationship between head and tail vertex
In addition to mandatory properties, each vertex or edge can also hold a set of custom properties. These properties can be defined by users, which can make vertices and edges appear similar to documents. In the table below, you can find a comparison between the graph model, the relational data model, and the OrientDB graph model:
Relational Model | Graph Model | OrientDB Graph Model |
---|---|---|
Table | Vertex and Edge Class | Class that extends "V" (for Vertex) and "E" (for Edges) |
Row | Vertex | Vertex |
Column | Vertex and Edge property | Vertex and Edge property |
Relationship | Edge | Edge |
The Key/Value Model
This is the simplest model of the three. Everything in the database can be reached by a key, where the values can be simple and complex types. OrientDB supports Documents and Graph Elements as values allowing for a richer model, than what you would normally find in the classic Key/Value model. The classic Key/Value model provides "buckets" to group key/value pairs in different containers. The most classic use cases of the Key/Value Model are:
- POST the value as payload of the HTTP call ->
/<bucket>/<key>
- GET the value as payload from the HTTP call ->
/<bucket>/<key>
- DELETE the value by Key, by calling the HTTP call ->
/<bucket>/<key>
The table below illustrates the comparison between the relational model, the Key/Value model, and the OrientDB Key/Value model:
Relational Model | Key/Value Model | OrientDB Key/Value Model |
---|---|---|
Table | Bucket | Class or Cluster |
Row | Key/Value pair | Document |
Column | not available | Document field or Vertex/Edge property |
Relationship | not available | Link |
The Object Model
This model has been inherited by Object Oriented programming and supports Inheritance between types (sub-types extends the super-types), Polymorphism when you refer to a base class and Direct binding from/to Objects used in programming languages.
The table below illustrates the comparison between the relational model, the Object model, and the OrientDB Object model:
Relational Model | Object Model | OrientDB Object Model |
---|---|---|
Table | Class | Class or Cluster |
Row | Object | Document or Vertex |
Column | Object property | Document field or Vertex/Edge property |
Relationship | Pointer | Link |
Multi Tenant
There are at many ways to build multi-tenant applications on top of OrientDB, in this page we are going to analyze pros and cons of three of the most used approaches.
One database per tenant
With this solution, each tenant is a database. The OrientDB server allows to host multiple databases.
Pros:
- Easy to use: to create/drop a new tenant, simply create/drop the database
- Easy to scale up by moving the database on different servers
Cons:
- Hard to create reports and analytics cross tenant, it requires to execute the same query against all the databases
Specific clusters per tenant
With this solution, each tenant is stored in one or more clusters of the same database. For example, the class Product
could have the following clusters: Product_ClientA
, Product_ClientB
, Product_ClientC
. In this way a query against a specific cluster will be used to retrieve data from one tenant only. Example to retrieve all the products of 2016, ordered by the most recent, only for the tenant "ClientC": select * from cluster:Product_ClientC where date >= '2016-01-01' order by date desc
.
Instead, a query against the class Product
will return a cross-tenant result. Example: select * from Product where date >= '2016-01-01' order by date desc
.
Pros:
- Easy to create reports and analytics cross tenant, because it's just one database
- It's possible to scale up on multiple servers by using sharding
Cons:
- The maximum number of clusters per database is 32,768. To bypass this limitation, use multiple databases
- No security out of the box, it's entirely up to the application to isolate access between tenants
Use Partitioned Graphs
By using the OrientDB's record level security, it's possible to have multiple partitions of graphs accessible by different users. For more information look at Partitioned Graphs.
Pros:
- Easy to create reports and analytics cross tenant, because it's just one database
Cons:
- Performance: record level security has an overhead, specially with queries
- Scales worse than "One database per tenant" solution, because the database will end up to be much bigger
- Sharding is not possible because records of multiple tenants are mixed on the same clusters
- Hard to guarantee a predictable level of service for all the users, because users connect to a tenant impact the other tenants
Basic Concepts
Record
The smallest unit that you can load from and store in the database. Records come in four types:
- Documents
- Blobs
- Vertices
- Edges
A Record is the smallest unit that can be loaded from and stored into the database. A record can be a Document, a Blob a Vertex or even an Edge.
Document
The Document is the most flexible record type available in OrientDB. Documents are softly typed and are defined by schema classes with defined constraints, but you can also use them in a schema-less mode too.
Documents handle fields in a flexible manner. You can easily import and export them in JSON format. For example,
{
"name" : "Jay",
"surname" : "Miner",
"job" : "Developer",
"creations" : [
{
"name" : "Amiga 1000",
"company" : "Commodore Inc."
}, {
"name" : "Amiga 500",
"company" : "Commodore Inc."
}
]
}
For Documents, OrientDB also supports complex relationships. From the perspective of developers, this can be understood as a persistent Map<String,Object>
.
BLOB
In addition to the Document record type, OrientDB can also load and store binary data. The BLOB record type was called RecordBytes
before OrientDB v2.2.
Vertex
In Graph databases, the most basic unit of data is the node, which in OrientDB is called a vertex. The Vertex stores information for the database. There is a separate record type called the Edge that connects one vertex to another.
Vertices are also documents. This means they can contain embedded records and arbitrary properties.
Edge
In Graph databases, an arc is the connection between two nodes, which in OrientDB is called an edge. Edges are bidirectional and can only connect two vertices.
Edges can be regular or lightweight. The Regular Edge saves as a Document, while the Lightweight Edge does not. For an understanding of the differences between these, see Lightweight Edges.
For more information on connecting vertices in general, see Relationships, below.
Record ID
When OrientDB generates a record, it auto-assigns a unique unit identifier, called a Record ID, or RID. The syntax for the Record ID is the pound sign with the cluster identifier and the position. The format is like this:
#<cluster>:<position>
.
-
Cluster Identifier: This number indicates the cluster to which the record belongs. Positive numbers in the cluster identifier indicate persistent records. Negative numbers indicate temporary records, such as those that appear in result-sets for queries that use projections.
-
Position: This number defines the absolute position of the record in the cluster.
NOTE: The prefix character
#
is mandatory to recognize a Record ID.
Records never lose their identifiers unless they are deleted. When deleted, OrientDB never recycles identifiers. Additionally, you can access records directly through their Record ID's. For this reason, you don't need to create a field to serve as the primary key, as you do in Relational databases.
Record Version
Records maintain their own version number, which increments on each update. In optimistic transactions, OrientDB checks the version in order to avoid conflicts at commit time.
Class
The concept of the Class is taken from the Object Oriented Programming paradigm. In OrientDB, classes define records. It is closest to the concept of a table in Relational databases.
Classes can be schema-less, schema-full or a mix. They can inherit from other classes, creating a tree of classes. Inheritance, in this context, means that a sub-class extends a parent class, inheriting all of its attributes.
Each class has its own clusters (data files). A non-abstract class (see below) must have at least one cluster defined, which functions as its default cluster. But, a class can support multiple clusters. When you execute a query against a class, it automatically propagates to all clusters that are part of the class. When you create a new record, OrientDB selects the cluster to store it in using a configurable strategy.
When you create a new class, by default, OrientDB creates new persistent clusters with the same name as the class, in lowercase, suffixed with underscore and an integer. As a default, OrientDB creates as many clusters per class as many cores (processors) the host machine has.
Eg. for class Person
, OrientDB will create clusters person
, person_1
, person_2
and so on so forth.
Abstract Class
The concept of an Abstract Class is one familiar to Object-Oriented programming. In OrientDB, this feature has been available since version 1.2.0. Abstract classes are classes used as the foundation for defining other classes. They are also classes that cannot have instances. For more information on how to create an abstract class, see CREATE CLASS.
This concept is essential to Object Orientation, without the typical spamming of the database with always empty, auto-created clusters.
For more information on Abstract Class as a concept, see Abstract Type and Abstract Methods and Classes
Class vs. Cluster in Queries
The combination of classes and clusters is very powerful and has a number of use cases. Consider an example where you create a class Invoice
, with two clusters invoice2015
and invoice2016
. You can query all invoices using the class as a target with SELECT
.
orientdb> SELECT FROM Invoice
In addition to this, you can filter the result-set by year. The class Invoice
includes a year
field, you can filter it through the WHERE
clause.
orientdb> SELECT FROM Invoice WHERE year = 2012
You can also query specific objects from a single cluster. By splitting the class Invoice
across multiple clusters, (that is, one per year), you can optimize the query by narrowing the potential result-set.
orientdb> SELECT FROM CLUSTER:invoice2012
Due to the optimization, this query runs significantly faster, because OrientDB can narrow the search to the targeted cluster.
Cluster
Where classes provide you with a logical framework for organizing data, clusters provide physical or in-memory space in which OrientDB actually stores the data. It is comparable to the collection in Document databases and the table in Relational databases.
When you create a new class, the CREATE CLASS
process also creates physical clusters that serve as the default location in which to store data for that class. OrientDB forms the cluster names using the class name, with all lower case letters. Beginning with version 2.2, OrientDB creates additional clusters for each class, (one for each CPU core on the server), to improve performance of parallelism.
For more information, see the Clusters Tutorial.
Materialized View
A materialized view is a persistent object that contains the result of a query. In terms of SQL querying, it can be considered as the equivalent of a class, that means that it can be used as a target for queries.
A materialized view can be configured to be read-only or updatable. Updating a record of a materialized view results in the update of the original record (ie. the record from which the view raw was created).
Views can have indexes, like normal classes.
Relationships
OrientDB supports two kinds of relationships: referenced and embedded. It can manage relationships in a schema-full or schema-less scenario.
Referenced Relationships
In Relational databases, tables are linked through JOIN
commands, which can prove costly on computing resources. OrientDB manges relationships natively without computing JOIN
's. Instead, it stores direct links to the target objects of the relationship. This boosts the load speed for the entire graph of connected objects, such as in Graph and Object database systems.
For example
customer
Record A -------------> Record B
CLASS=Invoice CLASS=Customer
RID=5:23 RID=10:2
Here, record A
contains the reference to record B
in the property customer
. Note that both records are reachable by other records, given that they have a Record ID.
With the Graph API, Edges are represented with two links stored on both vertices to handle the bidirectional relationship.
1:1 and 1:n Referenced Relationships
OrientDB expresses relationships of these kinds using links of the LINK
type.
1:n and n:n Referenced Relationships
OrientDB expresses relationships of these kinds using a collection of links, such as:
LINKLIST
An ordered list of links.LINKSET
An unordered set of links, which does not accept duplicates.LINKMAP
An ordered map of links, withString
as the key type. Duplicates keys are not accepted.
With the Graph API, Edges connect only two vertices. This means that 1:n relationships are not allowed. To specify a 1:n relationship with graphs, create multiple edges.
Embedded Relationships
When using Embedded relationships, OrientDB stores the relationship within the record that embeds it. These relationships are stronger than Reference relationships. You can represent it as a UML Composition relationship.
Embedded records do not have their own Record ID, given that you can't directly reference it through other records. It is only accessible through the container record.
In the event that you delete the container record, the embedded record is also deleted. For example,
address
Record A <>----------> Record B
CLASS=Account CLASS=Address
RID=5:23 NO RID!
Here, record A
contains the entirety of record B
in the property address
. You can reach record B
only by traversing the container record. For example,
orientdb> SELECT FROM Account WHERE address.city = 'Rome'
1:1 and n:1 Embedded Relationships
OrientDB expresses relationships of these kinds using the EMBEDDED
type.
1:n and n:n Embedded Relationships
OrientDB expresses relationships of these kinds using a collection of links, such as:
EMBEDDEDLIST
An ordered list of records.EMBEDDEDSET
An unordered set of records, that doesn't accept duplicates.EMBEDDEDMAP
An ordered map of records as the value and a string as the key, it doesn't accept duplicate keys.
Inverse Relationships
In OrientDB, all Edges in the Graph model are bidirectional. This differs from the Document model, where relationships are always unidirectional, requiring the developer to maintain data integrity. In addition, OrientDB automatically maintains the consistency of all bidirectional relationships.
Database
The database is an interface to access the real Storage. IT understands high-level concepts such as queries, schemas, metadata, indices and so on. OrientDB also provides multiple database types. For more information on these types, see Database Types.
Each server or Java VM can handle multiple database instances, but the database name must be unique. You can't manage two databases at the same time, even if they are in different directories. To handle this case, use the $
dollar character as a separator instead of the /
slash character. OrientDB binds the entire name, so it becomes unique, but at the file system level it converts $
with /
, allowing multiple databases with the same name in different paths. For example,
test$customers -> test/customers
production$customers = production/customers
Database URL
OrientDB uses its own URL format, of engine and database name as <engine>:<db-name>
.
Engine | Description | Example |
---|---|---|
plocal | This engine writes to the file system to store data. There is a LOG of changes to restore the storage in case of a crash. | plocal:/temp/databases/petshop/petshop |
memory | Open a database completely in memory | memory:petshop |
remote | The storage will be opened via a remote network connection. It requires an OrientDB Server up and running. In this mode, the database is shared among multiple clients. Syntax: remote:<server>:[<port>]/db-name . The port is optional and defaults to 2424. | remote:localhost/petshop |
Database Usage
You must always close the database once you finish working on it.
NOTE: OrientDB automatically closes all opened databases, when the process dies gracefully (not by killing it by force). This is assured if the Operating System allows a graceful shutdown.
OrientDB Programs
When you install OrientDB on your system, there are a number of applications included in the package. You can find them in the $ORIENTDB_HOME/bin/
directory. These applications include the OrientDB Server, the Console used to connect to the server and various other programs and administrative utilities.
OrientDB Server
OrientDB provides several methods of starting and managing the Server process. These include individual scripts that initialize the Java Virtual Machine and start OServer and systemd and initscripts that allow you to launch and manage the Server using the systemctl
or service
utilities on Linux.
Programs
There are two scripts that you can use to start the OrientDB Server directly. Each script has two builds: a .sh
shell script for use on Linux distributions and a .bat
batch file for use on Windows.
Script | Description |
---|---|
dserver | Used to start the OrientDB Server in distributed environments |
server | Used to start a standalone OrientDB Server |
Process Management
Where the above programs make it easy to start the Server on your desktop for experimentation and similar tasks, it is not very convenient for use on a production host. In addition to the Server startup scripts, OrientDB also ships with a systemd service file and an initscript, which you can install on Linux distributions to start, stop and restart the OrientDB Server using either systemctl
or service
utilities.
For more information, see OrientDB Server Process.
OrientDB Server - server
The OrientDB Server is a Java application that manages database operations, retrieval and storage. The shell script server
provides a convenient command-line interface to use in configuring and launching the OrientDB Server.
OrientDB provides a server.sh
shell script for starting the server on Linux operating systems and a server.bat
batch file for starting it on Windows.
Configuration Environmental Variables
OrientDB manages most of its configuration options through various XML configuration files. However, there are a handful of environmental variables that it utilizes when starting the Server. When the script runs, it check these variables to determine whether any value has been set on them. In the event that it finds none, it instead sets a sensible default value.
On Linux, you can set the environmental variable using the export
command:
$ export ORIENTDB_HOME=/opt/orientdb/3.2.33/
The following environmental variables are used by server.sh
:
Variable Name | Default | Description |
---|---|---|
$CONFIG_FILE | $ORIENTDB_HOME/config/orientdb-server-config.xml | XML configuration file for the server |
$JAVA_HOME | $PATH/java | Path to Java home directory, used when you have multiple installations of Java available on the host |
$JAVA_OPTS_SCRIPT | -Djna.nosys=true -XX:+HeapDumpOnOutOfMemoryError -Djava.awt.headless=true -Dfile.encoding=UTF8 -Drhino.opt.level=9 | Defines |
$ORIENTDB_HOME | current working directory | OrientDB installation directory |
$ORIENTDB_LOG_CONF | $ORIENTDB_HOME/config/orientdb-server-log.properties | Path to Java properties file you want to use |
$ORIENTDB_OPTS_MEMORY | -Xms2G -Xmx2G | Memory options, defaults to 2GB of heap |
$ORIENTDB_PID | $ORIENTDB_HOME/bin/orient.pid | Path to the process id (.pid ) file |
$ORIENTDB_SETTINGS | Default settings for OrientDB | |
$ORIENTDB_WWW_PATH | $ORIENTDB_HOME/www/ | Path to web resources, defaults to the interface to launch OrientDB Studio |
The server.sh
script also accepts argument. If one of the arguments is debug
, the script enables the following Java options -Xdebug -Xrunjdwp:transport=dt_socket,server=y,suspend=n,address=1044
. Any additional arguments are passed directly to the java
program.
OrientDB Server - dserver.sh
In situations where an individual host is insufficient for your storage or performance needs, you can run OrientDB on multiple hosts distributing the load between them. Where the server
script starts OrientDB as an individual Server, the dserver
script does the same, while configuring Server to operate in a distributed environment.
OrientDB provides a dserver.sh
shell script for Linux operating systems and dserver.bat
batch file for the Windows operating system.
Configuration Environmental Variables
OrientDB manages most of its configuration options through various XML configuration files. However, there are a handful of environmental variables that it utilizes when starting the Server. When the script runs, it check these variables to determine whether any value has been set on them. In the event that it finds none, it instead sets a sensible default value.
On Linux, you can set the environmental variable using the export
command:
$ export ORIENTDB_HOME=/opt/orientdb/3.2.33/
The following environmental variables are used by server.sh
:
Variable Name | Default | Description |
---|---|---|
$CONFIG_FILE | $ORIENTDB_HOME/config/orientdb-server-config.xml | XML configuration file for the server |
$JAVA_HOME | $PATH/java | Path to Java home directory, used when you have multiple installations of Java available on the host |
$JAVA_OPTS_SCRIPT | -Djna.nosys=true -XX:+HeapDumpOnOutOfMemoryError -Djava.awt.headless=true -Dfile.encoding=UTF8 -Drhino.opt.level=9 | Defines |
$ORIENTDB_HOME | current working directory | OrientDB installation directory |
$ORIENTDB_LOG_CONF | $ORIENTDB_HOME/config/orientdb-server-log.properties | Path to Java properties file you want to use |
$ORIENTDB_OPTS_MEMORY | -Xms2G -Xmx2G | Memory options, defaults to 2GB of heap |
$ORIENTDB_PID | $ORIENTDB_HOME/bin/orient.pid | Path to the process id (.pid ) file |
$ORIENTDB_SETTINGS | Default settings for OrientDB | |
$ORIENTDB_WWW_PATH | $ORIENTDB_HOME/www/ | Path to web resources, defaults to the interface to launch OrientDB Studio |
The dserver.sh
script also accepts argument. If one of the arguments is debug
, the script enables the following Java options -Xdebug -Xrunjdwp:transport=dt_socket,server=y,suspend=n,address=1044
. Any additional arguments are passed directly to the java
program.
OrientDB Server Process
Using the server
or dserver
shell scripts or batch files you can manually start the OrientDB Server. While this is often sufficient for testing or experimentation, however, in production environments it is preferable to user process and service management tools, like init and systemd.
Service Installation
In order to manage OrientDB through process or service management tools, you first need to install it on your system. The bin/
directory of your OrientDB installation contains an orientdb.sh
script that you need to modify and install at /etc/init.d/orientdb
.
For instance,
# cp $ORIENTDB_HOME/bin/orientdb.sh /etc/init.d/orientdb
Once you have the file in place, edit it and modify the two script variables at the top of the file.
Variable | Description |
---|---|
ORIENTDB_HOME | Defines the path where you've installed OrientDB |
ORIENTDB_USER | Defines the system user you want the script to switch to before it starts the server |
Service File
In cases where your host uses systemd for service and process management, you also need to copy the orientdb.service
file.
# cp $ORIENTDB_HOME/bin/orientdb.service /etc/systemd/system
Once this is done, edit the service file to update values on the following fields:
Field | Description |
---|---|
User | Set to the user you want to switch to when running the OrientDB Server |
Group | Set to the group you want to switch to when running the OrientDB Server |
ExecStart | Set to the path to orientdb init script copied above, /etc/init.d/orientdb |
Usage
On operating systems that use init for process management, copying the OrientDB script and setting the variables is enough to manage it using the service
command. On hosts that utilize systemd, you also need to copy over the orientdb.service
file.
Usage with init
When working with a host that uses init, you can manage the OrientDB Server using the following commands:
To start the OrientDB Server:
# service orientdb start
To stop the OrientDB Server
# service orientdb stop
To start OrientDB when the server boots:
# service orientdb enable
Usage with systemd
When working with a host that uses systemd, you can manage the OrientDB Server using the following commands:
To start the OrientDB Server:
# systemctl start orientdb
To stop the OrientDB Server
# systemctl stop orientdb
To start OrientDB when the server boots:
# systemctl enable orientdb
OrientDB Client Programs
OrientDB ships with one client utility: the Console. The Console is an interactive text-based interface for administering OrientDB Server and databases.
Console
OrientDB provides a Console Tool, which is a Java application that connects to and operates on OrientDB databases and Server instances.
Console Modes
There are two modes available to you, while executing commands through the OrientDB Console: interactive mode and batch mode.
Interactive Mode
By default, the Console starts in interactive mode. In this mode, the Console loads to an orientdb>
prompt. From there you can execute commands and SQL statements as you might expect in any other database console.
You can launch the console in interactive mode by executing the console.sh
for Linux OS systems or console.bat
for Windows systems in the bin
directory of your OrientDB installation. Note that running this file requires execution permissions.
$cd $ORIENTDB_HOME/bin
$./console.sh
OrientDB console v.X.X.X (build 0) www.orientdb.com Type 'HELP' to display all the commands supported. Installing extensions for GREMLIN language v.X.X.X orientdb>
From here, you can begin running SQL statements or commands. For a list of these commands, see commands.
Batch mode
When the Console runs in batch mode, it takes commands as arguments on the command-line or as a text file and executes the commands in that file in order. Use the same console.sh
or console.bat
file found in bin
at the OrientDB installation directory.
-
Command-line: To execute commands in batch mode from the command line, pass the commands you want to run in a string, separated by a semicolon.
$
$ORIENTDB_HOME/bin/console.sh "CONNECT REMOTE:localhost/demo;SELECT FROM Profile"
-
Script Commands: In addition to entering the commands as a string on the command-line, you can also save the commands to a text file as a semicolon-separated list.
$
vim commands.txt
CONNECT REMOTE:localhost/demo;SELECT FROM Profile
$$ORIENTDB_HOME/bin/console.sh commands.txt
Ignoring Errors
When running commands in batch mode, you can tell the console to ignore errors, allowing the script to continue the execution, with the ignoreErrors
setting.
$vim commands.txt
SET ignoreErrors TRUE
Enabling Echo
Regardless of whether you call the commands as an argument or through a file, when you run console commands in batch mode, you may also need to display them as they execute. You can enable this feature using the echo
setting, near the start of your commands list.
$vim commands.txt
SET echo TRUE
Enabling Date in prompt
Starting from v2.2.9, to enable the date in the prompt, set the variable promptDateFormat
with the date format following the SimpleDateFormat specs.
orientdb {db=test1}> set promptDateFormat "yyy-MM-dd hh:mm:ss.sss"
orientdb {db=test1 (2016-08-26 09:34:12.012)}>
Console Commands
OrientDB implements a number of SQL statements and commands that are available through the Console. In the event that you need information while working in the console, you can access it using either the HELP
or ?
command.
Command | Description |
---|---|
BACKUP DATABASE | Backup a database |
BROWSE CLASS | Browses all the records of a class |
BROWSE CLUSTER | Browses all the records of a cluster |
CLASSES | Displays all the configured classes |
CLUSTER STATUS | Displays the status of distributed cluster of servers |
CLUSTERS | Displays all the configured clusters |
CONFIG | Displays the configuration where the opened database is located (local or remote) |
CONFIG GET | Returns a configuration value |
CONFIG SET | Set a configuration value |
CONNECT | Connects to a database (deprecated see OPEN) |
CONNECT ENV | Connects to an OrientDB server/environment |
CREATE DATABASE | Creates a new database |
DECLARE INTENT | Declares an intent |
DICTIONARY KEYS | Displays all the keys in the database dictionary |
DICTIONARY GET | Lookups for a record using the dictionary. If found set it as the current record |
DICTIONARY PUT | Inserts or modify an entry in the database dictionary. The entry is composed by key=String, value=record-id |
DICTIONARY REMOVE | Removes the association in the dictionary |
DISCONNECT | Disconnects from the current database |
DISPLAY RECORD | Displays current record's attributes |
DISPLAY RAW RECORD | Displays current record's raw format |
DROP DATABASE | Drop a database |
EXPORT DATABASE | Exports a database |
EXPORT RECORD | Exports a record in any of the supported format (i.e. json) |
FREEZE DATABASE | Freezes the database locking all the changes. Use this to raw backup. Once frozen it uses the RELEASE DATABASE to release it |
GET | Returns the value of a property |
IMPORT DATABASE | Imports a database previously exported |
INDEXES | Displays information about indexes |
INFO | Displays information about current status |
INFO CLASS | Displays information about a class |
JS | Executes a Javascript in the console |
JSS | Executes a Javascript in the server |
LIST DATABASES | List the available databases |
LIST CONNECTIONS | List the available connections |
LOAD RECORD | Loads a record in memory and set it as the current one |
LOAD SCRIPT | Loads and executes a sql script |
OPEN | Opens a database connection on the currently connected server/environment (see CONNECT ENV) |
PROFILER | Controls the Profiler |
PROPERTIES | Returns all the configured properties |
pwd | Display current path |
RELEASE DATABASE | Releases a Console Freeze Database database |
RELOAD RECORD | Reloads a record in memory and set it as the current one |
RELOAD SCHEMA | Reloads the schema |
RESTORE DATABASE | Restore a database |
SET | Changes the value of a property |
HELP | Prints this help |
EXIT | Closes the console |
SQL Commands |
Console - BACKUP
Executes a complete backup on the currently opened database. It then compresses the backup file using the ZIP algorithm. You can then restore a database from backups, using the RESTORE DATABASE
command. You can automate backups using the Automatic-Backup server plugin.
Backups and restores are similar to the EXPORT DATABASE
and IMPORT DATABASE
, but they offer better performance than these options.
NOTE: OrientDB Community Edition does not support backing up remote databases. OrientDB Enterprise Edition does support this feature. For more information on how to implement this with Enterprise Edition, see Remote Backups.
Syntax:
BACKUP DATABASE <output-file> [-incremental] [-compressionLevel=<compressionLevel>] [-bufferSize=<bufferSize>]
<output-file>
Defines the path to the backup file.-incremental
Option to execute an incremental backup. When enabled, it computes the data to backup as all new changes since the last backup. Available only in the OrientDB Enterprise Edition version 2.2 or later.- -
compressionLevel
Defines the level of compression for the backup file. Valid levels are0
to9
. The default is9
. Available in 1.7 or later. -bufferSize
Defines the compression buffer size. By default, this is set to 1MB. Available in 1.7 or later.
Permissions:
In order to enable a user to execute this command, you must add the permission of create
for the resource database.backup
to the database user.
Example:
- Backing up a database:
orientdb>CONNECT plocal:../databases/mydatabase admin admin
orientdb>BACKUP DATABASE /backups/mydb.zip
Backing current database to: database mydb.zip Backup executed in 0.52 seconds
Incremental Backup
Since version 2.2, OrientDB Enterprise Edition supports incremental backups. For more details see Incremental Backup and Restore
For more information, see:
Console - BROWSE CLASS
Displays all records associated with the given class.
Syntax:
BROWSE CLASS <class-name>
<class-name>
Defines the class for the records you want to display.
Permissions:
In order to enable a user to execute this command, you must add the permission of read
for the resource database.class.<class>
to the database user.
Example:
-
Browse records associated with the class
City
:orientdb>
BROWSE CLASS City
----+------+------------------- # | RID | NAME ----+------+------------------- 0 | -6:0 | Rome 1 | -6:1 | London 2 | -6:2 | Honolulu ----+------+-------------------
For more information on other commands, see Console Commands.
Console - BROWSE CLUSTER
Displays all records associated with the given cluster.
Syntax:
BROWSE CLUSTER <cluster-name>
<cluster-name>
Defines the cluster for the records you want to display.
Permissions:
In order to enable a user to execute this command, you must add the permission of read
for the resource database.cluster.<class>
to the database user.
Example:
-
Browse records associated with the cluster
City
:orientdb>
BROWSE CLUSTER City
----+------+------------------- # | RID | NAME ----+------+------------------- 0 | -6:0 | Rome 1 | -6:1 | London 2 | -6:2 | Honolulu ----+------+-------------------
For more information on other commands, see Console Commands.
Console - LIST CLASSES
Displays all configured classes in the current database.
Syntax:
-
Long Syntax:
LIST CLASSES
-
Short Syntax:
CLASSES
Example
-
List current classes in the database:
orientdb>
LIST CLASSES
CLASSES -------------+------+-------------+----------- NAME | ID | CLUSTERS | ELEMENTS -------------+------+-------------+----------- Person | 0 | person | 7 Animal | 1 | animal | 5 AnimalRace | 2 | AnimalRace | 0 AnimalType | 3 | AnimalType | 1 OrderItem | 4 | OrderItem | 0 Order | 5 | Order | 0 City | 6 | City | 3 -------------+------+-------------+----------- TOTAL 16 -----------------------------------------------
For more information on other commands, see Console Commands.
Console - CLUSTER STATUS
Displays the status of the cluster in distributed configuration.
Syntax:
CLUSTER STATUS
Example:
-
Display the status of the cluster:
orientdb>
CLUSTER STATUS
{ "localName": "_hzInstance_1_orientdb", "localId": "3735e690-9a7b-44d2-b4bc-27089da065e2", "members": [ { "id": "3735e690-9a7b-44d2-b4bc-27089da065e2", "name": "node1", "startedOn": "2015-05-14 17:06:40:418", "listeners": [ { "protocol": "ONetworkProtocolBinary", "listen": "10.3.15.55:2424" }, { "protocol": "ONetworkProtocolHttpDb", "listen": "10.3.15.55:2480" } ], "databases": [] } ] }
For more information on other commands, see Console Commands.
Console - LIST CLUSTERS
Displays all configured clusters in the current database.
Syntax:
-
Long Syntax:
LIST CLUSTERS
-
Short Syntax:
CLUSTERS
Example:
-
List current clusters on database:
orientdb>
LIST CLUSTERS
CLUSTERS -------------+------+-----------+----------- NAME | ID | TYPE | ELEMENTS -------------+------+-----------+----------- metadata | 0 | Physical | 11 index | 1 | Physical | 0 default | 2 | Physical | 779 csv | 3 | Physical | 1000 binary | 4 | Physical | 1001 person | 5 | Physical | 7 animal | 6 | Physical | 5 animalrace | -2 | Logical | 0 animaltype | -3 | Logical | 1 orderitem | -4 | Logical | 0 order | -5 | Logical | 0 city | -6 | Logical | 3 -------------+------+-----------+----------- TOTAL 2807 --------------------------------------------
For information on creating new clusters in the current database, see the
CREATE CLUSTER
command. For more information on other commands, see Console Commands.
Console - LIST SERVERS
Displays all active servers connected within a cluster.
This command was introduced in OrientDB version 2.2.
Syntax:
LIST SERVERS
Example:
-
List the servers currently connected to the cluster:
orientdb>
LIST SERVERS
CONFIGURED SERVERS -+----+------+-----------+-------------+-----------+-----------+-----------+----------+--------- #|Name|Status|Connections|StartedOn |Binary |HTTP |UsedMemory |FreeMemory|MaxMemory -+----+------+-----------+-------------+-----------+-----------+-----------+----------+--------- 0|no2 |ONLINE|0 |2015-10-30...|192.168.0.6|192.168.0.6|80MB(8.80%)|215MB(23%)|910MB 1|no1 |ONLINE|0 |2015-10-30...|192.168.0.6|192.168.0.6|90MB(2.49%)|195MB(5%) |3.5GB -+----+------+-----------+-------------+-----------+-----------+-----------+----------+--------- -
Use the
DISPLAY
command to show information on a specific server:orientdb>
DISPLAY 0
-------------+------------------------------ Name | Value -------------+------------------------------ Name | node2 Status | ONLINE Connections | 0 StartedOn | Fri Oct 30 21:41:07 CDT 2015 Binary | 192.168.0.6:2425 HTTP | 192.168.0.6:2481 UsedMemory | 80,16MB (8,80%) FreeMemory | 215,34MB (23,65%) MaxMemory | 910,50MB -------------+------------------------------
For more information on other commands, see Console Commands.
Console - LIST SERVER USERS
This feature was introduced in OrientDB version 2.2.
Displays all configured users on the server. In order to display the users, the current system user that is running the console must have permissions to read the $ORIENTDB_HOME/config/orientdb-server-config.xml
configuration file. For more information, see OrientDB Server Security.
Syntax:
LIST SERVER USERS
Example:
-
List configured users on a server:
orientdb>
LIST SERVER USERS
SERVER USERS - 'root', permissions: * - 'guest', permissions: connect,server.listDatabases,server.dblist
For more information, see
For more information on other console commands, see Console Commands.
Console - CHECK DATABASE
Checks the integrity of a database. In the case the database contains graphs, their consistency is checked. To repair a database, use Repair Database Command.
Syntax
CHECK DATABASE [--skip-graph] [-v]
[--skip-graph]
Skips the check of the graph[-v]
Verbose mode
Examples
- Check a graph database:
orientdb> CHECK DATABASE
Check of graph 'plocal:/temp/testdb' is started ...
Scanning 1 edges (skipEdges=0)...
+ found corrupted edge E#17:0{out:#9:0,in:#11:0,test:true} v2 because incoming vertex (#11:0) does not contain the edge
Scanning edges completed
Scanning 710 vertices...
+ found corrupted vertex V#10:0{in_:[#17:0],name:Marko} v2 the edge should be removed from property in_ (ridbag)
Scanning vertices completed
Check of graph 'plocal:/temp/testdb' completed in 0 secs
scannedEdges.....: 1
edgesToRemove....: 1
scannedVertices..: 710
scannedLinks.....: 2
linksToRemove....: 1
verticesToRepair.: 0
Check of storage completed in 296ms. without errors.
For more information on other commands, see Console Commands.
Console - CONFIG
Displays the configuration information on the current database, as well as whether it is local or remote.
Syntax
CONFIG
Examples
-
Display the configuration of the current database:
orientdb>
CONFIG
REMOTE SERVER CONFIGURATION: +------------------------------------+--------------------------------+ | NAME | VALUE | +------------------------------------+--------------------------------+ | treemap.lazyUpdates | 300 | | db.cache.enabled | false | | file.mmap.forceRetry | 5 | | treemap.optimizeEntryPointsFactor | 1.0 | | storage.keepOpen | true | | treemap.loadFactor | 0.7 | | file.mmap.maxMemory | 110000000 | | network.http.maxLength | 10000 | | storage.cache.size | 5000 | | treemap.nodePageSize | 1024 | | ... | ... | | treemap.entryPoints | 30 | +------------------------------------+--------------------------------+
You can change configuration variables displayed here using the
CONFIG SET
command. To display the value set to one configuration variable, use theCONFIG GET
command.For more information on other commands, see Console Commands.
Console - CONFIG GET
Displays the value of the requested configuration variable.
Syntax
CONFIG GET <config-variable>
<config-variable>
Defines the configuration variable you want to query.
Examples
-
Display the value to the
tx.log.fileType
configuration variable:orientdb>
CONFIG GET tx.log.fileType
Remote configuration: tx.log.fileType = classic
You can display all configuration variables using the
CONFIG
command. To change the values, use theCONFIG SET
command.For more information on other commands, see Config Commands.
Console - CONFIG SET
Updates a configuration variable to the given value.
Syntax
CONFIG SET <config-variable> <config-value>
<config-variable>
Defines the configuration variable you want to change.<config-value>
Defines the value you want to set.
Example
-
Display the current value for
tx.autoRetry
:orientdb>
CONFIG GET tx.autoRetry
Remote configuration: tx.autoRetry = 1Change the
tx.autoRetry
value to5
:orientdb>
CONFIG SET tx.autoRetry 5
Remote configuration value changed correctly.Display new value:
orientdb>
CONFIG GET tx.autoRetry
Remote configuration: tx.autoRetry = 5
You can display all configuration variables with the
CONFIG
command. You can view the current value on a configuration variable using theCONFIG GET
command.For more information on other commands, see Console Commands
Console - CONNECT
(deprecated, see CONNECT ENV and OPEN)
Opens a database or connects to a server.
Syntax
CONNECT <database-url> <user> <password>
<database-url>
Defines the URL of the database you want to connect to. It uses the format<mode>:<path>
<mode>
Defines the mode you want to use in connecting to the database. It can beplocal
orremote
.<path>
Defines the path to the database.
<user>
Defines the user you want to connect to the database with.<password>
Defines the password needed to connect to the database, with the defined user.
Examples:
-
Connect to a local database as the user
admin
, loading it directly into the console:orientdb>
connect plocal:../databases/GratefulDeadConcerts admin my_admin_password
Connecting to database [plocal:../databases/GratefulDeadConcerts]...OK -
Connect to a remote database:
orientdb>
connect remote:192.168.1.1/GratefulDeadConcerts admin my_admin_password
Connecting to database [remote:192.168.1.1/GratefulDeadConcerts]...OK
For more information on other commands, see Console Commands.
search: keywords: ['console', 'command', 'connection', 'CONNECT']
Console - CONNECT ENV
(since v 3.2)
Connects the console to an OrientDB server or to an embedded environment.
Syntax
CONNECT ENV <env-url> [<user> <password>]
<env-url>
defines the URL of an OrientDB server/environment. It uses the format<mode>:<path>
<mode>
Defines the mode you want to use in connecting to the database. It can beplocal
orremote
.<path>
Defines the path to the database.
<user>
Defines the user you want to connect with.<password>
Defines the password needed to connect, with the defined user. User and password are needed to execute authenticated operations on the environment (eg. create a database on a remote server); they are not needed for operations that do not require an authentication or that require different credentials (eg. to open a database you can pass the database user and password, see OPEN)
After the console is connected to an environment, it allows to execute operations and Server-Level commands on it.
Examples:
-
Connect the console to a local (embedded) environment as the user
root
, loading it directly into the console:orientdb>
connect env plocal:../databases/
-
Connect to a remote server:
orientdb>
connect env remote:192.168.1.1 root root
For more information on other commands, see Console Commands.
Console - CREATE DATABASE
WARNING: OrientDB v 3.2 includes some important changes to DB creation; in particular, default database users/passwords are not created anymore. The console command
CREATE DATABASE
had a specific behaviour that relied on default users: it created the database and THEN connected to the DB using admin/admin. This is not possible anymore, so we had to change the syntax AND the semantics of the command
Creates and connects to a new database.
Syntax
CREATE DATABASE <database-name> <storage-type> [users ( (<username> identified by <password> role <rolename>)* )]
<database-name>
Defines the name of the database you want to create.<storage-type>
Defines the storage type that you want to use. You can choose betweenPLOCAL
andMEMORY
.<username>
The name of a user to create<password>
The password of this newly created user<rolename>
The role name of this user
Examples
-
Create a local database
demo
without default users:orientdb>
CONNECT ENV embedded:/my/databases/dir root root
orientdb>CREATE DATABASE demo plocal
Database created successfully. -
Create a remote database
demo
without default users:orientdb>
CONNECT ENV remote:localhost root root
orientdb>CREATE DATABASE demo plocal
Database created successfully. -
Create a remote database
demo
with an "admin" and a "reader" user:orientdb>
CONNECT ENV remote:localhost root root
orientdb>CREATE DATABASE demo plocal users (admin identified by 'MyAdminPassword' role admin, reader identified by 'MyReaderPw' role reader)
Database created successfully.
Old Syntax
(still compatible in v 3.2)
CREATE DATABASE <database-url> [<user> <password> <storage-type> [<db-type>]] [-restore=<backup-path>]
<database-url>
Defines the URL of the database you want to connect to. It uses the format<mode>:<path>
<mode>
Defines the mode you want to use in connecting to the database. It can bePLOCAL
orREMOTE
.<path>
Defines the path to the database.
<user>
Defines the user you want to connect to the database with.<password>
Defines the password needed to connect to the database, with the defined user.<storage-type>
Defines the storage type that you want to use. You can choose betweenPLOCAL
andMEMORY
.<db-type>
Defines the database type. You can choose betweenGRAPH
andDOCUMENT
. The default isGRAPH
.
Examples
-
Create a local database
demo
:orientdb>
CREATE DATABASE PLOCAL:/usr/local/orientdb/databases/demo
Creating database [plocal:/usr/local/orientdb/databases/demo]... Connecting to database [plocal:/usr/local/orientdb/databases/demo]...OK Database created successfully. Current database is: plocal:/usr/local/orientdb/databases/demo orientdb {db=demo}> -
Create a remote database
trick
:orientdb>
CREATE DATABASE REMOTE:192.168.1.1/trick root E30DD873203AAA245952278B4306D94E423CF91D569881B7CAD7D0B6D1A20CE9 PLOCAL
Creating database [remote:192.168.1.1/trick ]... Connecting to database [remote:192.168.1.1/trick ]...OK Database created successfully. Current database is: remote:192.168.1.1/trick orientdb {db=trick}>
To create a static database to use from the server, see
Server pre-configured storage types
.To remove a database, see
DROP DATABASE
. To change database configurations after creation, seeALTER DATABASE
.For more information on other commands, see Console Commands.
Incremental restore option
You can execute an incremental restore at creation time through the option -restore
specifying as value the path where your backup is placed. Let's suppose we want create a new fresh database "mydb" and restore data from a backup, located in /tmp/backup
, performed from another database in one shot. In this case we can type:
orientdb> create database remote:localhost/mydb root root plocal graph -restore=/tmp/backup
Creating database [remote:localhost/mydb] using the storage type [plocal]...
Connecting to database [remote:localhost/mydb] with user 'admin'...OK
Database created successfully.
Current database is: remote:localhost/mydb
For further details on incremental backup and restore you can refer to the page Incremental Backup and Restore.
Console - DECLARE INTENT
Declares an intent for the current database. Intents allow you to tell the database what you want to do.
Syntax
DECLARE INTENT <intent-name>
<intent-name>
Defines the name of the intent. OrientDB supports three intents:NULL
Removes the current intent.MASSIVEINSERT
MASSIVEREAD
Examples
-
Declare an intent for a massive insert:
orientdb>
DECLARE INTENT MASSIVEINSERT
-
After the insert, clear the intent:
orientdb>
DECLARE INTENT NULL
For more information on other commands, see Console Commands.
Console - DICTIONARY GET
Displays the value of the requested key, loaded from the database dictionary.
Syntax
DICTIONARY GET <key>
<key>
Defines the key you want to access.
Example
-
In a dictionary of U.S. presidents, display the entry for Barack Obama:
orientdb>
DICTIONARY GET obama
------------------------------------------------------------------------- Class: Person id: 5:4 v.1 ------------------------------------------------------------------------- parent: null children : [Person@5:5{parent:Person@5:4,children:null,name:Malia Ann, surname:Obama,city:null}, Person@5:6{parent:Person@5:4, children:null,name:Natasha,surname:Obama,city:null}] name : Barack surname : Obama city : City@-6:2{name:Honolulu} -------------------------------------------------------------------------
You can display all keys stored in a database using the
DICTIONARY KEYS
command. For more information on indexes, see Indexes.
For more information on other commands, see Console Commands.
Console - DICTIONARY KEYS
Displays all the keys stored in the database dictionary.
Syntax
DICTIONARY KEYS
Example
-
Display all the keys stored in the database dictionary:
orientdb>
DICTIONARY KEYS
Found 4 keys: #0: key-148 #1: key-147 #2: key-146 #3: key-145
To load the records associated with these keys, use the
DICTIONARY GET
command. For more information on indexes, see Indexes.For more information on other commands, see Console Commands.
Console - DICTIONARY PUT
Binds a record to a key in the dictionary database, making it accessible to the DICTIONARY GET
command.
Syntax
DICTIONARY PUT <key> <record-id>
<key>
Defines the key you want to bind.<record-id>
Defines the ID for the record you want to bind to the key.
Example
-
In the database dictionary of U.S. presidents, bind the record for Barack Obama to the key
obama
:orientdb>
DICTIONARY PUT obama 5:4
------------------------------------------------------------------------ Class: Person id: 5:4 v.1 ------------------------------------------------------------------------ parent : null children : [Person@5:5{parent:Person@5:4,children:null,name:Malia Ann, surname:Obama,city:null}, Person@5:6{parent:Person@5:4, children:null,name:Natasha,surname:Obama,city:null}] name : Barack surname : Obama city : City@-6:2{name:Honolulu} ------------------------------------------------------------------------ The entry obama=5:4 has been inserted in the database dictionary
To see all the keys stored in the database dictionary, use the
DICTIONARY KEYS
command. For more information on dictionaries and indexes, see Indexes.For more information on other commands, see Console Commands.
Console - DICTIONARY REMOVE
Removes the association from the database dictionary.
Syntax
DICTIONARY REMOVE <key>
<key>
Defines the key that you want to remove.
Example
-
In a database dictionary of U.S. presidents, remove the key for Barack Obama:
orientdb>
DICTIONARY REMOVE obama
Entry removed from the dictionary. Last value of entry was: ------------------------------------------------------------------------ Class: Person id: 5:4 v.1 ------------------------------------------------------------------------ parent : null children : [Person@5:5{parent:Person@5:4,children:null,name:Malia Ann, surname:Obama,city:null}, Person@5:6{parent:Person@5:4, children:null,name:Natasha,surname:Obama,city:null}] name : Barack surname : Obama city : City@-6:2{name:Honolulu} ------------------------------------------------------------------------
You can display information for all keys stored in the database dictionary using the
DICTIONARY KEY
command. For more information on dictionaries and indexes, see Indexes.For more information on other commands, see Console Commands.
Console - DISCONNECT
Closes the currently opened database.
Syntax
DISCONNECT
Example
-
Disconnect from the current database:
orientdb>
DISCONNECT
Disconnecting from the database [../databases/petshop/petshop]...OK
To connect to a database, see
CONNECT
. For more information on other commands, see Console Commands.
Console - DISPLAYS RECORD
Displays details on the given record from the last returned result-set.
Syntax
DISPLAY RECORD <record-number>
<record-number>
Defines the relative position of the record in the last result-set.
Example
-
Query the database on the class
Person
to generate a result-set:orientdb>
SELECT FROM Person
---+-----+--------+----------+-----------+-----------+------ # | RID | PARENT | CHILDREN | NAME | SURNAME | City ---+-----+--------+----------+-----------+-----------+------ 0 | 5:0 | null | null | Giuseppe | Garibaldi | -6:0 1 | 5:1 | 5:0 | null | Napoleon | Bonaparte | -6:0 2 | 5:2 | 5:3 | null | Nicholas | Churchill | -6:1 3 | 5:3 | 5:2 | null | Winston | Churchill | -6:1 4 | 5:4 | null | [2] | Barack | Obama | -6:2 5 | 5:5 | 5:4 | null | Malia Ann | Obama | null 6 | 5:6 | 5:4 | null | Natasha | Obama | null ---+-----+--------+----------+-----------+-----------+------ 7 item(s) found. Query executed in 0.038 sec(s). -
With the result-set ready, display record number four in the result-set, (for Malia Ann Obama):
orientdb>
DISPLAY RECORD 5
------------------------------------------------------------------------ Class: Person id: 5:5 v.0 ------------------------------------------------------------------------ parent : Person@5:4{parent:null,children:[Person@5:5, Person@5:6], name:Barack,surname:Obama,city:City@-6:2} children : null name : Malia Ann surname : Obama city : null ------------------------------------------------------------------------
For more information on other commands, see Console Commands.
Console - DISPLAYS RAW RECORD
Displays details on the given record from the last returned result-set in a binary format.
Syntax
DISPLAY RAW RECORD <record-number>
<record-number>
Defines the relative position of the record in the last result-set.
Example
-
Query the database on the class
V
to generate a result-set:orientdb {db=GratefulDeadConcerts}>
SELECT song_type, name, performances FROM V LIMIT 6
-----+-------+--------+----------+-------------------------+-------------- # | @RID | @CLASS | song_type | name | performances -----+-------+--------+----------+-------------------------+-------------- 0 | #9:1 | V | cover | HEY BO DIDDLEY | 5 1 | #9:2 | V | cover | IM A MAN | 1 2 | #9:3 | V | cover | NOT FADE AWAY | 531 3 | #9:4 | V | original | BERTHA | 394 4 | #9:5 | V | cover | GOING DOWN THE ROAD... | 293 5 | #9:6 | V | cover | MONA | 1 6 | #9:7 | V | null | Bo_Diddley | null -----+-------+--------+-----------+------------------------+------------- LIMIT EXCEEDED: resultset contains more items not displayed (limit=6) 6 item(s) found. Query executed in 0.136 sec(s). -
Display raw record on the song "Hey Bo Diddley" from the result-set:
orientdb {db=GratefulDeadConcerts}>
DISPLAY RAW RECORD 0
Raw record content. The size is 292 bytes, while settings force to print first 150 bytes: ^@^BV^Rsong_type^@^@^@^G^Hname^@^@^@^G^Htype^@^@^@ ^G^Xperformances^@^@^@ ^A^^out_followed_by^@^@^@ ^V^\out_written)by^@^@^@ ^V^Vout_sung_by^@^@^@ ^V^\in_followed_by^@^@^@ ^V^@civer^\HEY Bo D
For more information on other commands available, see Console Commands.
Console - DROP DATABASE
Removes a database completely. If the database is open and a database name not given, it removes the current database.
Syntax
DROP DATABASE [<database-name> <server-username> <server-user-password>]
<database-name
Defines the database you want to drop. By default it uses the current database, if it's open.<server-username>
Defines the server user. This user must have the privileges to drop the database.<server-user-password>
Defines the password for the server user.
NOTE: When you drop a database, it deletes the database and all records, caches and schema information it contains. Unless you have made backups, there is no way to restore the database after you drop it.
Examples
-
Remove the current local database:
orientdb>
DROP DATABASE
-
Remove the database
demo
at localhost:orientdb>
DROP DATABASE REMOTE:localhost/demo root root_password
You can create a new database using the
CREATE DATABASE
command. To make changes to an existing database, use theALTER DATABASE
command.
For more information on other commands, please refer to Console Commands and SQL Commands.
Console - DROP SERVER USER
Removes a user from the server. In order to do so, the current system user running the Console, must have permissions to write to the $ORIENTDB_HOME/config/orientdb-server-config.xmL
configuration file.
Syntax
DROP SERVER USER <user-name>
<user-name>
Defines the user you want to drop.
NOTE: For more information on server users, see OrientDB Server Security.
This feature was introduced in version 2.2.
Example
-
Remove the user
editor
from the Server:orientdb>
DROP SERVER USER editor
Server user 'editor' dropped correctly
To view the current server users, see the
LIST SERVER USERS
command. To create or update a server user, see theSET SERVER USER
command.
For more information on other commands, see Console Commands.
Console - EXPORT
Exports the current database to a file. OrientDB uses a JSON-based Export Format. By default, it compresses the file using the GZIP algorithm.
With the IMPORT
command, this allows you to migrate the database between different versions of OrientDB without losing data.
If you receive an error about the database version, export the database using the same version of OrientDB that has generated the database.
Bear in mind, exporting a database browses it, rather than locking it. While this does mean that concurrent operations can execute during the export, it also means that you cannot create an exact replica of the database at the point when the command is issued. In the event that you need to create a snapshot, use the BACKUP
command.
You can restore a database from an export using the IMPORT
.
NOTE: While the export format is JSON, there are some constraints in the field order. Editing this file or adjusting its indentation may cause imports to fail.
Syntax
By default, this command exports the full database. Use its options to disable the parts you don't need to export.
EXPORT DATABASE <output-file>
[-excludeAll]
[-includeClass=<class-name>*]
[-excludeClass=<class-name>*]
[-includeCluster=<cluster-name>*]
[-excludeCluster=<cluster-name>*]
[-includeInfo=<true|false>]
[-includeClusterDefinitions=<true|false>]
[-includeSchema=<true|false>]
[-includeSecurity=<true|false>]
[-includeRecords=<true|false>]
[-includeIndexDefinitions=<true|false>]
[-includeManualIndexes=<true|false>]
[-compressionLevel=<0-9>]
[-compressionBuffer=<bufferSize>]
<output-file>
Defines the path to the output file.-excludeAll
Sets the export to exclude everything not otherwise included through command options-includeClass
Export includes certain classes, specifically those defined by a space-separated list.In case you specify multiple class names, you have to wrap the list between quotes, eg.-includeClass="Foo Bar Baz"
-excludeClass
Export excludes certain classes, specifically those defined by a space-separated list.-includeCluster
Export includes certain clusters, specifically those defined by a space-separated list.-excludeCluster
Export excludes certain clusters, specifically those defined by a space-separated list.-includeInfo
Defines whether the export includes database information.-includeClusterDefinitions
Defines whether the export includes cluster definitions.-includeSchema
Defines whether the export includes the database schema.-includeSecurity
Defines whether the export includes database security parameters.-includeRecords
Defines whether the export includes record contents.-includeIndexDefinitions
Defines whether the export includes the database index definitions.-includeManualIndexes
Defines whether the export includes manual index contents.-compressionLevel
Defines the compression level to use on the export, in a range between0
(no compression) and9
(maximum compression). The default is1
. (Feature introduced in version 1.7.6.)-compressionBuffer
Defines the compression buffer size in bytes to use in compression. The default is 16kb. (Feature introduced in version 1.7.6.)
Examples
-
Export the current database, including everything:
orientdb>
EXPORT DATABASE C:\temp\petshop.export
Exporting current database to: C:\temp\petshop.export... Exporting database info...OK Exporting dictionary...OK Exporting schema...OK Exporting clusters... - Exporting cluster 'metadata' (records=11) -> ...........OK - Exporting cluster 'index' (records=0) -> OK - Exporting cluster 'default' (records=779) -> OK - Exporting cluster 'csv' (records=1000) -> OK - Exporting cluster 'binary' (records=1001) -> OK - Exporting cluster 'person' (records=7) -> OK - Exporting cluster 'animal' (records=5) -> OK - Exporting cluster 'animalrace' (records=0) -> OK - Exporting cluster 'animaltype' (records=1) -> OK - Exporting cluster 'orderitem' (records=0) -> OK - Exporting cluster 'order' (records=0) -> OK - Exporting cluster 'city' (records=3) -> OK Export of database completed. -
Export the current database, including only its functions:
orientdb>
EXPORT DATABASE functions.gz -includeClass=OFunction -includeInfo=FALSE -includeClusterDefinitions=FALSE -includeSchema=FALSE -includeIndexDefinitions=FALSE -includeManualIndexes=FALSE
-
Alternatively, you can simplify the above by excluding all, then including only those features that you need. For instance, export the current database, including only the schema:
orientdb>
EXPORT DATABASE schema.gz -excludeALL -includeSchema=TRUE
For more information on backups and restores, imports and exports, see the following commands:
as well as the following pages:
- Export File Format
ODatabaseExport
Java ClassFor more information on other commands, see Console Commands.
Console - EXPORT RECORD
Exports the current record, using the requested format. In the event that you give a format that OrientDB does not support, it provides a list of supported formats.
Syntax
EXPORT RECORD <format>
<format>
Defines the export format you want to use.
Examples
-
Use
SELECT
to create a record for export:orientdb>
SELECT name, surname, parent, children, city FROM Person WHERE name='Barack' AND surname='Obama'
---+-----+--------+---------+--------+------------+------ # | RID | name | surname | parent | children | city ---+-----+--------+---------+--------+------------+------ 0 | 5:4 | Barack | Obama | null | [5:5, 5:6] | -6:2 ---+-----+--------+---------+--------+------------+------ -
Export JSON data from this record:
orientdb>
EXPORT RECORD JSON
{ 'name': 'Barack', 'surname': 'Obama', 'parent': null, 'children': [5:5, 5:6], 'city': -6:2 } -
Use a bad format value to determine what export formats are available on your database:
orientdb>
EXPORT RECORD GIBBERISH
ERROR: Format 'GIBBERISH' was not found. Supported formats are: - json - ORecordDocument2csv
For more information on other commands, see Console Commands.
Console - FREEZE DATABASE
Flushes all cached content to disk and restricts permitted operations to read commands. With the exception of reads, none of the commands made on a frozen database execute. It remains in this state until you run the RELEASE
command.
Executing this command requires server administration rights. You can only execute it on remote databases. If you would like to freeze or release a local database, use the ODatabase.freeze()
and ODatabase.release()
methods directly through the OrientDB API.
You may find this command useful in the event that you would like to perform backups on a live database. To do so, freeze the database, perform a file system snapshot, then release the database. You can now copy the snapshot anywhere you want.
This works best when the backup doesn't take very long to run.
Syntax
FREEZE DATABASE
Example
-
Freezes the current database:
orientdb>
FREEZE DATABASE
To unfreeze a database, use the
RELEASE DATABASE
command.
For more information on other commands, please refer to Console Commands and SQL Commands.
Console - GET
Returns the value of the requested property.
Syntax
GET <property-name>
<property-name>
Defines the name of the property.
Example
-
Find the default limit on your database:
orientdb>
GET LIMIT
limit = 20
To display all available properties configured on your database, use the
PROPERTIES
command.
For more information on other commands, see Console Commands.
Console - IMPORT
Imports an exported database into the current one open. Import process doesn't lock the database, so any concurrent operations are allowed, but they could interfere in the import process causing errors.
The input file must use the JSON Export Format, as generated by the EXPORT
command. By default, this file is compressed using the GZIP algorithm.
With EXPORT
, this command allows you to migrate between releases without losing data, by exporting data from the old version and importing it into the new version.
Syntax
IMPORT DATABASE <input-file> [-format = <format>]
[-preserveClusterIDs = <true|false>]
[-deleteRIDMapping = <true|false>]
[-merge = <true|false>]
[-migrateLinks = <true|false>]
[-rebuildIndexes = <true|false>]
<format>
Is the input file format. If not specified, OrientDB tries to recognize it. The available formats are (since v2.2.8):- orientdb, the OrientDB export file format
- graphml, for Graph XML
- graphson, for Graph JSON
<input-file>
Defines the path to the file you want to import.-preserveClusterIDs
Defines whether you want to preserve cluster ID's during the import. When turned off, the import creates temporary cluster ID's, which can sometimes fail. This option is only valid with PLocal storage.-deleteRIDMapping
Defines whether you want to preserve the dictionary index used by the import to map old RIDs to new RIDs. The index name is___exportImportRIDMap
and you could use in your application. By default the index is removed after the import.-merge
Defines whether you want to merge the import with the data already in the current database. When turned off, the default, the import overwrites current data, with the exception of security classes, (ORole
,OUser
,OIdentity
), which it always preserves. This feature was introduced in version 1.6.1.-migrateLinks
Defines whether you want to migrate links after the import. When enabled, this updates all references from the old links to the new Record ID's. By default, it is enabled. Advisable that you only turn it off when merging and you're certain no other existent records link to those you're importing. This feature was introduced in version 1.6.1.-rebuildIndexes
Defines whether you want to rebuild indexes after the import. By default, it does. You can set it to false to speed up the import, but do so only when you're certain the import doesn't affect indexes. This feature was introduced in version 1.6.1.
Example
-
Import the database
petshop.export
:orientdb>
IMPORT DATABASE C:/temp/petshop.export -preserveClusterIDs=true
Importing records... - Imported records into the cluster 'internal': 5 records - Imported records into the cluster 'index': 4 records - Imported records into the cluster 'default': 1022 records - Imported records into the cluster 'orole': 3 records - Imported records into the cluster 'ouser': 3 records - Imported records into the cluster 'csv': 100 records - Imported records into the cluster 'binary': 101 records - Imported records into the cluster 'account': 1005 records - Imported records into the cluster 'company': 9 records - Imported records into the cluster 'profile': 9 records - Imported records into the cluster 'whiz': 1000 records - Imported records into the cluster 'address': 164 records - Imported records into the cluster 'city': 55 records - Imported records into the cluster 'country': 55 records - Imported records into the cluster 'animalrace': 3 records - Imported records into the cluster 'ographvertex': 102 records - Imported records into the cluster 'ographedge': 101 records - Imported records into the cluster 'graphcar': 1 records
For more information on backups, restores, and exports, see:
BACKUP
,RESTORE
andEXPORT
commands, and theODatabaseImport
Java class. For the JSON format, see Export File Format.
For more information on other commands, see Console Commands.
Troubleshooting
Validation Errors
Occasionally, you may encounter validation errors during imports, usually shown as an OValidationException
exception. Beginning with version 2.2, you can disable validation at the database-level using the ALTER DATABASE
command, to allow the import to go through.
-
Disable validation for the current database:
orientdb>
ALTER DATABASE validation false
-
Import the exported database:
orientdb>
IMPORT DATABASE /path/to/my_data.export -preserveClusterIDs=TRUE
-
Re-enable validation:
orientdb>
ALTER DATABASE validation true
Cluster ID's
During imports you may occasionally encounter an error that reads: Imported cluster 'XXX' has id=6 different from the original: 5
. Typically occurs in databases that were created in much older versions of OrientDB. You can correct it using the DROP CLASS
on the class ORIDs
, then attempting the import again.
-
Import the database:
orientdb>
IMPORT DATABASE /path/to/old_data.export
Importing records...
-
Creating cluster 'company'...Error on database import happened just before line 16, column 52 com.orientechnologies.orient.core.exception.OConfigurationException: Imported cluster 'company has id=6 different from the original: 5 at com.orientechnologies.orient.core.db.tool.ODatabaseImport.importClusters( ODatabaseImport.java:500) at com.orientechnologies.orient.core.db.tool.ODatabaseIMport.importDatabase( ODatabaseImport.java:121)
-
Drop the
ORIDs
class:orientdb>
DROP CLASS ORIDs
-
Import the database:
orientdb>
IMPORT DATABASE /path/to/old_data.export
The database now imports without error.
Console - INDEXES
Displays all indexes in the current database.
Syntax
INDEXES
Example
-
Display indexes in the current database:
orientdb {db=GratefulDeadConcerts}>
INDEXES
INDEXES --------------+------------+-------+--------+--------- NAME | TYPE | CLASS | FIELDS | RECORDS --------------+------------+-------+--------+--------- dictionary | DICTIONARY | | | 0 Group.Grp_Id | UNIQUE | Group | Grp_Id | 1 ORole.name | UNIQUE | ORole | name | 3 OUser.name | UNIQUE | OUser | name | 4 --------------+------------+----------------+--------- TOTAL = 4 8 ------------------------------------------------------
For more information on other commands, see Console Commands.
Console - INFO
Displays all information on the current database.
Syntax
INFO
Example
-
Display information on database
petshop
:orientdb {db=petshop}>
INFO
Current database: ../databases/petshop/petshop CLUSTERS: ------------+------+----------+---------- NAME | ID | TYPE | ELEMENTS ------------+------+----------+---------- metadata | 0 | Physical | 11 index | 1 | Physical | 0 default | 2 | Physical | 779 csv | 3 | Physical | 1000 binary | 4 | Physical | 1001 person | 5 | Physical | 7 animal | 6 | Physical | 5 animalrace | -2 | Logical | 0 animaltype | -3 | Logical | 1 orderitem | -4 | Logical | 0 order | -5 | Logical | 0 city | -6 | Logical | 3 ------------+------+----------+---------- TOTAL 2807 ----------------------------------------- CLASSES: ------------+----+------------+---------- NAME | ID | CLUSTERS | ELEMENTS ------------+----+------------+---------- Person | 0 | person | 7 Animal | 1 | animal | 5 AnimalRace | 2 | AnimalRace | 0 AnimalType | 3 | AnimalType | 1 OrderItem | 4 | OrderItem | 0 Order | 5 | Order | 0 City | 6 | City | 3 ------------+----+------------+---------- TOTAL 16 -----------------------------------------
For more information on other commands, see Console Commands.
Console - INFO CLASS
Displays all information on given class.
Syntax
INFO CLASS <class-name>
<class-name>
Defines what class you want information on.
Example
-
Display information on class
Profile
orientdb>
INFO CLASS Profile
Default cluster......: profile (id=10) Supported cluster ids: [10] Properties: --------+----+----------+-----------+---------+-----------+----------+-----+---- NAME | ID | TYPE | LINK TYPE | INDEX | MANDATORY | NOT NULL | MIN | MAX --------+----+----------+-----------+---------+-----------+----------+-----+---- nick | 3 | STRING | null | | false | false | 3 | 30 name | 2 | STRING | null |NOTUNIQUE| false | false | 3 | 30 surname| 1 | STRING | null | | false | false | 3 | 30 ... | | ... | ... | ... | ... | ... |... | ... photo | 0 | TRANSIENT| null | | false | false | | --------+----+----------+-----------+---------+-----------+----------+-----+----
For more information on other commands, see Console Commands.
Console - INFO PROPERTY
Displays all information on the given property.
Syntax
INFO PROPERTY <class-name>.<property-name>
<class-name>
Defines the class to which the property belongs.<property-name>
Defines the property you want information on.
Example
-
Display information on the property
name
in the classOUser
:orientdb>
INFO PROPERTY OUser.name
PROPERTY 'OUser.name' Type.................: STRING Mandatory............: true Not null.............: true Read only............: false Default value........: null Minimum value........: null Maximum value........: null REGEXP...............: null Collate..............: {OCaseInsensitiveCollate : name = ci} Linked class.........: null Linked type..........: null INDEXES (1 altogether) --------------------+------------ NAME | PROPERTIES --------------------+------------ OUser.name | name --------------------+------------
For more information on other commands, see Console Commands.
Console - JS
Executes commands in the Javascript language from the Console. Look also Javascript Command.
Syntax
JS <commands>
<commands>
Defines the commands you want to execute.
Interactive Mode
You can execute a command in just one line (JS print('Hello World!')
) or enable the interactive input by just executing JS
and then typing the Javascript expression as multi-line inputs. It does not execute the command until you type end
. Bear in mind, the end
here is case-sensitive.
Examples
-
Execute a query and display the result:
orientdb>
js
[Started multi-line command. Type just 'end' to finish and execute.] orientdb>var r = db.query('select from ouser');
orientdb>for(var i=0;i
orientdb> print( r[i] );
orientdb>}
orientdb>end
OUser#5:0{roles:[1],status:ACTIVE,password:{PBKDF2WithHmacSHA256}C08CE0F5160EA4050B8F10EDBB86F06EB0A2EE82DF73A340:BC1B6040727C1E11E3A961A1B2A49615C96938710AF17ADD:65536,name:admin} v1 OUser#5:1{name:reader,password:{PBKDF2WithHmacSHA256}41EF9B675430D215E0970AFDEB735899B6665DF44A29FE98:5BC48B2D20752B12B5E32BE1F22C6C85FF7CCBEFB318B826:65536,status:ACTIVE,roles:[1]} v1 OUser#5:2{name:writer,password:{PBKDF2WithHmacSHA256}FA0AD7301EA2DB371355EB2855D63F4802F13858116AB82E:18B8077E1E63A45DB0A3347F91E03E4D2218EA16E5100105:65536,status:ACTIVE,roles:[1]} v1 Client side script executed in 0.142000 sec(s). Value returned is: null
For more information on the Javascript execution, see Javascript Command. For more information on other commands, see Console Commands.
Console - JSS
Executes commands on OrientDB Server in the Javascript language from the Console. Look also Javascript Command.
Syntax
JSS <commands>
<commands>
Defines the commands you want to execute.
Interactive Mode
You can execute a command in just one line (JSS print('Hello World!')
) or enable the interactive input by just executing JSS
and then typing the Javascript expression as multi-line inputs. It does not execute the command until you type end
. Bear in mind, the end
here is case-sensitive.
Examples
-
Execute a query and display the result:
orientdb>
jss
[Started multi-line command. Type just 'end' to finish and execute.] orientdb>var r = db.query('select from ouser');
orientdb>for(var i=0;i
orientdb> print( r[i] );
orientdb>}
orientdb>end
Server side script executed in 0.146000 sec(s). Value returned is: nullIn this case the output will be displayed on the server console.
For more information on the Javascript execution, see Javascript Command. For more information on other commands, see Console Commands.
Console - LIST DATABASES
Displays all databases hosted on the current server. Note that this command requires you connect to the OrientDB Server.
Syntax
LIST DATABASES
Example
-
Connect to the server:
orientdb>
CONNECT REMOTE:localhost admin admin_password
-
List the databases hosted on the server:
orientdb {server=remote:localhost/}>
LIST DATABASES
Found 4 databases: * ESA (plocal) * Napster (plocal) * Homeland (plocal) * GratefulDeadConcerts (plocal)
For more information on other commands, see Console Commands.
Console - LIST CONNECTIONS
Displays all active connections to the OrientDB Server. Command introduced in version 2.2. The connections as per server, so you should connect to the server, not to the database.
Syntax
LIST CONNECTIONS
Permissions
In order to enable a user to execute this command, you must add "server.info"
as resource to the server user.
Example
-
List the current connections to the OrientDB Server:
orientdb {server=remote:localhost/}>
LIST CONNECTIONS
---+----+--------------+------+-------------------+--------+-----+--------+-------- # | ID |REMOTE_ADDRESS|PROTOC|LAST_OPERATION_ON |DATABASE|USER |COMMAND |TOT_REQS ---+----+--------------+------+-------------------+--------+-----+--------+-------- 0 | 17 |/127.0.0.1 |binary|2015-10-12 19:22:34|- |- |info | 1 1 | 16 |/127.0.0.1 |binary|1970-01-01 01:00:00|- |- |- | 0 5 | 1 |/127.0.0.1 |http |1970-01-01 00:59:59|pokec |admin|Listen | 32 ---+----+--------------+------+-------------------+--------+-----+--------+--------
For more information on other commands, see Console Commands.
Console - LOAD RECORD
Loads a record the given Record ID from the current database.
Syntax
LOAD RECORD <record-id>
<record-id
Defines the Record ID of the record you want to load.
In the event that you don't have a Record ID, execute a query to find the one that you want.
Example
-
Load the record for
#5:5
:orientdb>
LOAD RECORD #5:5
-------------------------------------------------------------------------------- Class: Person id: #5:5 v.0 -------------------------------------------------------------------------------- parent : Person@5:4{parent:null,children:[Person@5:5, Person@5:6],name:Barack, surname:Obama,city:City@-6:2} children : null name : Malia Ann surname : Obama city : null --------------------------------------------------------------------------------
For more information on other commands, see Console Commands.
Console - LOAD SCRIPT
(from 2.2.18)
Loads a sql script from the given path and executes it.
Syntax
LOAD SCRIPT <script path>
Example
-
Load a script from an absolute path:
orientdb>
LOAD SCRIPT /path/to/scripts/data.osql
-
Launch the console in batch mode and load script to a remote database:
$
$ORIENTDB_HOME/bin/console.sh "CONNECT REMOTE:localhost/demo;LOAD SCRIPT /path/to/scripts/data.osql"
For more information on other commands, see Console Commands.
search: keywords: ['console', 'command', 'connection', 'CONNECT']
Console - OPEN
(since v 3.2)
Opens a database.
Syntax
OPEN <database-name> <user> <password>
<database-name>
Defines the name of the database you want to connect to. The console has to be connected to a server/environment (see CONNECT ENV ) and the database has to be present in that environment<user>
Defines the user you want to connect to the database with.<password>
Defines the password needed to connect to the database, with the defined user.
Examples:
-
Connect to a local database as the user
admin
, loading it directly into the console:orientdb>
connect env plocal:../databases/
orientdb>open demodb admin my_admin_password
Connecting to database [plocal:../databases/demodb]...OK -
Connect to a remote database:
orientdb>
connect env remote:192.168.1.1
orientdb>open demodb admin my_admin_password
Connecting to database [remote:192.168.1.1/demodb]...OK
For more information on other commands, see Console Commands.
Console - PROFILER
Controls the Profiler.
Syntax
PROFILER ON|OFF|DUMP|RESET
ON
Turn on the Profiler and begin recording.OFF
Turn off the Profiler and stop recording.DUMP
Dump the Profiler data.RESET
Reset the Profiler data.
Example
-
Turn the Profiler on:
orientdb>
PROFILER ON
Profiler is ON now, use 'profiler off' to turn off. -
Dump Profiler data:
orientdb>
PROFILER DUMP
For more information on other commands, see Console Commands.
Console - PROPERTIES
Displays all configured properties.
Syntax
PROPERTIES
Example
-
List configured properties:
orientdb>
PROPERTIES
PROPERTIES: ------------------------+----------- NAME | VALUE ------------------------+----------- limit | 20 backupBufferSize | 1048576 backupCompressionLevel | 9 collectionMaxItems | 10 verbose | 2 width | 150 maxBinaryDisplay | 150 debug | false ignoreErrors | false ------------------------+-----------
To change a property value, use the
SET
command.For more information on other commands, see Console Commands.
Console - RELEASE DATABASE
Releases database from a frozen state, from where it only allows read operations back to normal mode. Execution requires server administration rights.
You may find this command useful in the event that you want to perform live database backups. Run the FREEZE DATABASE
command to take a snapshot, you can then copy the snapshot anywhere you want. Use such approach when you want to take short-term backups.
Syntax
RELEASE DATABASE
Example
-
Release the current database from a freeze:
orientdb>
RELEASE DATABASE
To freeze a database, see the
FREEZE DATABASE
command.For more information on other commands, please refer to Console Commands and SQL Commands.
Console - RELOAD RECORD
Reloads a record from the current database by its Record ID, ignoring the cache.
You may find this command useful in cases where external applications change the record and you need to see the latest update.
Syntax
RELOAD RECORD <record-id>
<record-id>
Defines the unique Record ID for the record you want to reload. If you don't have the Record ID, execute a query first.
Examples
-
Reload record with the ID of
5:5
:orientdb>
RELOAD RECORD 5:5
------------------------------------------------------------------------ Class: Person id: 5:5 v.0 ------------------------------------------------------------------------ parent : Person@5:4{parent:null,children:[Person@5:5, Person@5:6], name:Barack,surname:Obama,city:City@-6:2} children : null name : Malia Ann surname : Obama city : null ------------------------------------------------------------------------
For more information on other commands, see Console Commands.
Console - REPAIR DATABASE
Repairs a database. To check if a database needs to be repaired, you can use the Check Database Command.
Syntax
REPAIR DATABASE [--fix-graph [-skipVertices=<vertices>] [-skipEdges=<edges>]]
[--fix-links] [-v]]
[--fix-ridbags]
[--fix-bonsai]
[-v]
[--fix-graph]
Fixes the database as graph. All broken edges are removed and all corrupted vertices repaired. This mode takes the following optional parameters:-skipVertices=<vertices>
, where<vertices>
are the number of vertices to skip on repair.-skipEdges=<edges>
, where<edges>
are the number of edges to skip on repair.[--fix-links]
Fixes links. It removes any reference to not existent records. The optional[-v]
tells to print more information.[--fix-ridbags]
Fixes the ridbag structures only (collection of references).[--fix-bonsai]
Fixes the bonsai structures only (internal representation of trees)[-v]
Verbose mode
Examples
- Repair a graph database:
orientdb> REPAIR DATABASE --fix-graph
Repair of graph 'plocal:/temp/demo' is started ...
Scanning 26632523 edges (skipEdges=0)...
...
+ edges: scanned 100000, removed 0 (estimated remaining time 10 secs)
+ edges: scanned 200000, removed 0 (estimated remaining time 9 secs)
+ deleting corrupted edge friend#40:22044{out:#25:1429,in:#66:1,enabled:true} v7 because missing incoming vertex (#66:1)
...
Scanning edges completed
Scanning 32151775 vertices...
+ vertices: scanned 100000, repaired 0 (estimated remaining time 892 secs)
+ vertices: scanned 200000, repaired 0 (estimated remaining time 874 secs)
+ vertices: scanned 300000, repaired 0 (estimated remaining time 835 secs)
+ repaired corrupted vertex Account#25:961{out_friend:[],dateUpdated:Wed Aug 12 19:00:00 CDT 2015,createdOn:Wed Aug 12 19:00:00 CDT 2015} v4
...
+ vertices: scanned 32100000, repaired 47 (estimated remaining time 2 secs)
...
Scanning vertices completed
Repair of graph 'plocal:/temp/demo' completed in 2106 secs
scannedEdges.....: 1632523
removedEdges.....: 129
scannedVertices..: 32151775
scannedLinks.....: 53264852
removedLinks.....: 64
repairedVertices.: 47
For more information on other commands, see Console Commands.
Console - RESTORE DATABASE
Restores a database from a backup. It must be done against a new database. It does not support restores that merge with an existing database. If you need to backup and restore to an existing database, use the EXPORT DATABASE
and IMPORT DATABASE
commands.
OrientDB Enterprise Edition version 2.2 and major, support incremental backup.
To create a backup file to restore from, use the BACKUP DATABASE
command.
Syntax
RESTORE DATABASE <backup-file>|<incremental-backup-directory>
<backup-file>
Defines the database file you want to restore.<incremental-backup-directory>
Defines the database directory you want to restore from an incremental backup. Available only in OrientDB Enterprise Edition version 2.2 and major.
Permissions:
In order to enable a user to execute this command, you must add the permission of create
for the resource database.restore
to the database user.
Example of full restore
-
Create a new database to receive the restore:
orientdb>
CREATE DATABASE PLOCAL:/tmp/mydb
-
Restore the database from the
mydb.zip
backup file:orientdb {db=/tmp/mydb}>
RESTORE DATABASE /backups/mydb.zip
Example of incremental restore
This is available only in OrientDB Enterprise Edition version 2.2 and major.
-
Open a database to receive the restore:
orientdb>
CONNECT PLOCAL:/tmp/mydb
-
Restore the database from the
/backup
backup directory:orientdb {db=/tmp/mydb}>
RESTORE DATABASE /backup
For more information, see the
BACKUP DATABASE
,EXPORT DATABASE
,IMPORT DATABASE
commands. For more information on other commands, see Console Commands.
Console - SET
Changes the value of a property.
Syntax
SET <property-name> <property-value>
<property-name>
Defines the name of the property<property-value>
Defines the value you want to change the property to.
Example
-
Change the
LIMIT
property to one hundred:orientdb>
SET LIMIT 100
Previous value was: 20 limit = 100
To display all properties use the
PROPERTIES
command. To display the value of a particular property, use theGET
command.For more information on other commands, see Console Commands.
Console - SET SERVER USER
Creates a server user. If the server user already exists, it updates the password and permissions.
In order to create or modify the user, the current system user must have write permissions on the $ORIENTDB_HOME/config/orientdb-server-config.xml
configuration file.
Syntax
SET SERVER USER <user-name> <user-password> <user-permissions>
<user-name>
Defines the server username.<user-password>
Defines the password for the server user.<user-permissions>
Defines the permissions for the server user.
For more information on security, see OrientDB Server Security. Feature introduced in version 2.2.
Example
-
Create the server user
editor
, give it all permissions:orientdb>
SET SERVER USER editor my_password *
Server user 'editor' set correctly
To display all server users, see the
LIST SERVER USERS
command. To remove a server user, seeDROP SERVER USER
command.For more information on other commands, see Console Commands.
Custom Console Commands
In addition to the commands implemented by OrientDB, you can also develop custom commands to extend features in your particular implementation. To do this, edit the OConsoleDatabaseApp class and add to it a new method. There's an auto-discovery system in place that adds the new method to the available commands. To provide a description of the command, use annotations. The command name must follow the Java code convention of separating words using camel-case.
For instance, consider a case in which you might want to add a MOVE CLUSTER
command to the console:
@ConsoleCommand(description = "Move the physical location of cluster files")
public void moveCluster(
@ConsoleParameter(name = "cluster-name", description = "The name or the id of the cluster to remove") String iClusterName,
@ConsoleParameter(name = "target-path", description = "path of the new position where to move the cluster files") String iNewPath ) {
checkCurrentDatabase(); // THE DB MUST BE OPENED
System.out.println("Moving cluster '" + iClusterName + "' to path " + iNewPath + "...");
}
Once you have this code in place, MOVE CLUSTER
now appears in the listing of available commands shown by HELP
.
orientdb>HELP
AVAILABLE COMMANDS: * alter classAlter a class in the database schema * alter cluster Alter class in the database schema ... ... * move cluster Move the physical location of cluster files ... ... * help Print this help * exit Close the console orientdb> MOVE CLUSTER foo /temp
Moving cluster 'foo' to path /tmp...
In the event that you develop a custom command and find it especially useful in your deployment, you can contribute your code to the OrientDB Community!
OrientDB Administrative and Utility Programs
#OrientDB Data Import Programs
OrientDB Server Administration
OrientDB has a number of tools to make administration of the database easier. There is the Console, which allows you to run a large number of commands.
There is also the OrientDB Studio, which allows you to run queries and visually look at the graph.
OrientDB also offers several tools for the import and export of data, logging and trouble shooting, along with ETL tools.
All of OrientDB's administration facilities are aimed to make your usage of OrientDB as simple and as easy as possible.
For more information see:
Installation
OrientDB Community Edition is available as a binary package for download or as source code on GitHub. The Enterprise Edition is available as a binary package to all our Customers that purchased one of the available Subscriptions.
OrientDB prerequisites can be found here.
Binary Installation
OrientDB provides a pre-compiled binary package to install the database on your system. Depending on your operating system, this is a tarred or zipped package that contains all the relevant files you need to run OrientDB. For desktop installations, go to OrientDB Downloads and select the package that best suits your system.
On server installations, you can use the wget
utility:
$ wget https://repo1.maven.org/maven2/com/orientechnologies/orientdb-community/3.2.33/orientdb-community-3.2.33.tar.gz -O orientdb-community-3.2.33.tar.gz
Whether you use your web browser or wget
, unzip or extract the downloaded file into a directory convenient for your use, (for example, /opt/orientdb/
on Linux). This creates a directory called orientdb-community-3.2.33 with relevant files and scripts, which you will need to run OrientDB on your system.
Use Docker
If you have Docker installed in your computer, this is the easiest way to run OrientDB. From the command line type:
$ docker run -d --name orientdb -p 2424:2424 -p 2480:2480
-e ORIENTDB_ROOT_PASSWORD=root orientdb:latest
Where instead of "root", type the root's password you want to use.
More details on how to build your own OrientDB Docker image can be found here
Use Ansible
If you manage your servers through Ansible, you can use the following role : https://galaxy.ansible.com/migibert/orientdb which is highly customizable and allows you to deploy OrientDB as a standalone instance or multiple clusterized instances.
For using it, you can follow these steps :
Install the role
ansible-galaxy install migibert.orientdb
Create an Ansible inventory
Assuming you have one two servers with respective IPs fixed at 192.168.10.5 and 192.168.10.6, using ubuntu user.
[orientdb-servers]
192.168.20.5 ansible_ssh_user=ubuntu
192.168.20.6 ansible_ssh_user=ubuntu
Create an Ansible playbook
In this example, we provision a two node cluster using multicast discovery mode. Please note that this playbook assumes java is already installed on the machine so you should have one step before that install Java 8 on the servers
- hosts: orientdb-servers
become: yes
vars:
orientdb_version: 2.0.5
orientdb_enable_distributed: true
orientdb_distributed:
hazelcast_network_port: 2434
hazelcast_group: orientdb
hazelcast_password: orientdb
multicast_enabled: True
multicast_group: 235.1.1.1
multicast_port: 2434
tcp_enabled: False
tcp_members: []
orientdb_users:
- name: root
password: root
tasks:
- apt:
name: openjdk-8-jdk
state: present
roles:
- role: orientdb-role
Run the playbook
ansible-playbook -i inventory playbook.yml
Source Code Installation
For information on how to install OrientDB from source, please refer to this Section.
Post-installation Tasks
For desktop users installing the binary, OrientDB is now installed and can be run through shell scripts found in the package bin
directory of the installation. For servers, there are some additional steps that you need to take in order to manage the database server for OrientDB as a service. The procedure for this varies, depending on your operating system.
Upgrading
When the time comes to upgrade to a newer version of OrientDB, the methods vary depending on how you chose to install it in the first place. If you installed from binary downloads, repeat the download process above and update any symbolic links or shortcuts to point to the new directory.
For systems where OrientDB was built from source, pull down the latest source code and compile from source.
$ git pull origin master
$ mvn clean install
Bear in mind that when you build from source, you can switch branches to build different versions of OrientDB using Git. For example,
$ git checkout 2.2.x
$ mvn clean install
builds the 2.2.x
branch, instead of master
.
Building a single executable jar with OrientDB
OrientDB for internal components like engines, operators, factories uses Java SPI Service Provider Interface. That means that the jars of OrientDB are shipped with files in META-INF/services
that contains the implementation of components. Bear in mind that when building a single executable jar, you have to concatenate the content of files with the same name in different orientdb-*.jar . If you are using Maven Shade Plugin you can use Service Resource Transformer to do that.
Other Resources
To learn more about how to install OrientDB on specific environments, please refer to the guides below:
- Install with Docker
- Install with Ansible
- Install on Linux Ubuntu
- Install on JBoss AS
- Install on GlassFish
- Install on Ubuntu 12.04 VPS (DigitalOcean)
- Install on Vagrant
Prerequisites
Both editions of OrientDB run on any operating system that implements the Java Virtual machine (JVM) from v 8 to v 11, specifically the JDK. Examples of these include:
- Linux, all distributions, including ARM (Raspberry Pi, etc.)
- Mac OS X
- Microsoft Windows
OrientDB requires Java, version 8 to 11, of the JDK.
Note: In OSGi containers, OrientDB uses a
ConcurrentLinkedHashMap
implementation provided by concurrentlinkedhashmap to create the LRU based cache. This library actively uses the sun.misc package which is usually not exposed as a system package. To overcome this limitation you should add propertyorg.osgi.framework.system.packages.extra
with valuesun.misc
to your list of framework properties.It may be as simple as passing an argument to the VM starting the platform:
$ java -Dorg.osgi.framework.system.packages.extra=sun.misc
Installation from Source
In addition to downloading the binary packages, you also have the option of compiling OrientDB from the Community Edition source code, available on GitHub. This process requires that you install Git and Apache Maven on your system.
To compile OrientDB from source code, clone the Community Edition repository, then run Maven (mvn
) in the newly created directory:
$ git clone https://github.com/orientechnologies/orientdb
$ git checkout develop
$ cd orientdb
$ mvn clean install
It is possible to skip tests:
$ mvn clean install -DskipTests
The develop branch contains code for the next version of OrientDB. Stable versions are tagged on master branch.
For each maintained version OrientDB has its own hotfix
branch.
As the time of writing this notes, the state of branches is:
- develop: work in progress for next 3.0.x release (3.0.x-SNAPSHOT)
- 2.2.x: hot fix for next 2.2.x stable release (2.2.x-SNAPSHOT)
- 2.1.x: hot fix for next 2.1.x stable release (2.1.x-SNAPSHOT)
- 2.0.x: hot fix for next 2.0.x stable release (2.0.x-SNAPSHOT)
- last tag on master is 2.2.0
The build process installs all jars in the local maven repository and creates archives under the distribution
module inside the target
directory. At the time of writing, building from branch 2.1.x gave:
$ls -l distribution/target/
total 199920
1088 26 Jan 09:57 archive-tmp
102 26 Jan 09:57 databases
102 26 Jan 09:57 orientdb-community-2.2.1-SNAPSHOT.dir
48814386 26 Jan 09:57 orientdb-community-2.2.1-SNAPSHOT.tar.gz
53542231 26 Jan 09:58 orientdb-community-2.2.1-SNAPSHOT.zip
$
The directory orientdb-community-2.2.1-SNAPSHOT.dir
contains the OrientDB distribution uncompressed.
Take a look to Contribute to OrientDB if you want to be involved.
Update Permissions
For Linux, Mac OS X and UNIX-based operating system, you need to change the permissions on some of the files after compiling from source.
$ chmod 755 bin/*.sh
$ chmod -R 777 config
These commands update the execute permissions on files in the config/
directory and shell scripts in bin/
, ensuring that you can run the scripts or programs that you've compiled.
Install as Service on Unix/Linux
Following the installation guide above, whether you choose to download binaries or build from source, does not install OrientDB at a system-level. There are a few additional steps you need to take in order to manage the database system as a service.
OrientDB ships with a script, which allows you to manage the database server as a system-level daemon. You can find it in the bin/
path of your installation directory, (that is, at $ORIENTDB_HOME/bin/orientdb.sh
.
The script supports three parameters:
start
stop
status
Configuring the Script
In order to use the script on your system, you need to edit the file to define two variables: the path to the installation directory and the user you want to run the database server.
$ vi $ORIENTDB_HOME/bin/orientdb.sh
#!/bin/sh
# OrientDB service script
#
# Copyright (c) Orient Technologies LTD (http://www.orientechnologies.com)
# chkconfig: 2345 20 80
# description: OrientDb init script
# processname: orientdb.sh
# You have to SET the OrientDB installation directory here
ORIENTDB_DIR="YOUR_ORIENTDB_INSTALLATION_PATH"
ORIENTDB_USER="USER_YOU_WANT_ORIENTDB_RUN_WITH"
Edit the ORIENTDB_DIR
variable to indicate the installation directory. Edit the ORIENTDB_USER
variable to indicate the user you want to run the database server, (for instance, orientdb
).
Installing the Script
Different operating systems and Linux distributions have different procedures when it comes to managing system daemons, as well as the procedure for starting and stopping them during boot up and shutdown. Below are generic guides for init and systemd based unix systems as well Mac OS X. For more information, check the documentation for your particular system.
Installing for init
Many Unix-like operating systems such as FreeBSD, most older distributions of Linux, as well as current releases of Debian, Ubuntu and their derivatives use variations on SysV-style init for these processes. These are typically the systems that manage such processes using the service
command.
To install OrientDB as a service on an init-based unix or Linux system, copy the modified orientdb.sh
file from $ORIENTDB_HOME/bin
into /etc/init.d/
:
# cp $ORIENTDB_HOME/bin/orientdb.sh /etc/init.d/orientdb
Once this is done, you can start and stop OrientDB using the service
command:
# service orientdb start
Starting OrientDB server daemon...
Installing for systemd
Most newer releases of Linux, especially among the RPM-based distributions like Red Hat, Fedora, and CentOS, as well as future releases of Debian and Ubuntu use systemd for these processes. These are the systems that manage such processes using the systemctl
command.
The OrientDB's package contains a service descriptor file for systemd based distros. The orientdb.service
is placed in the bin
directory. To install OrientDB copy the orientdb.service
to/etc/systemd/system
directory (check this, may depend on distro). Edit the file:
# vi /etc/systemd/system/orientdb.service
#
# Copyright (c) OrientDB LTD (http://http://orientdb.com/)
#
[Unit]
Description=OrientDB Server
After=network.target
After=syslog.target
[Install]
WantedBy=multi-user.target
[Service]
User=ORIENTDB_USER
Group=ORIENTDB_GROUP
ExecStart=$ORIENTDB_HOME/bin/server.sh
Set the right user and group. You may want to use the absolute path instead of the environment variable $ORIENTDB_HOME
. Once this file is saved, you can start and stop the OrientDB server using the systemctl
command:
# systemctl start orientdb.service
Additionally, with the orientdb.service
file saved, you can set systemd to start the database server automatically during boot by issuing the enable
command:
# systemctl enable orientdb.service
Synchronizing state of orientdb.service with SysV init with /usr/lib/systemd/systemd-sysv-install...
Executing /usr/lib/systemd/systemd-sysv-install enable orientdb
Created symlink from /etc/systemd/system/multi-user.target.wants/orientdb.service to /etc/systemd/system/orientdb.service.
Installing for Mac OS X
Manual install
For Mac OS X:
- follow the steps described above, in the
Configuring the Script
section - create an alias to the OrientDB system daemon script and the console.
$ alias orientdb-server=/path/to/$ORIENTDB_HOME/bin/orientdb.sh
$ alias orientdb-console=/path/to/$ORIENTDB_HOME/bin/console.sh
You can now start the OrientDB database server using the following command:
$ orientdb-server start
Once the database starts, it is accessible through the console script.
$ orientdb-console
OrientDB console v.1.6 www.orientechnologies.com
Type 'HELP' to display all the commands supported.
orientdb>
Brew
OrientDB is available through brew.
$ brew install orientdb
The installation process gives an output similar to the following one:
...
==> Downloading https://orientdb.com/download.php?file=orientdb-community-<ORIENTDB_VERSION>.tar.gz
==> /usr/bin/nohup /usr/local/Cellar/orientdb/<ORIENTDB_VERSION>/libexec/bin/server.sh &
==> /usr/local/Cellar/orientdb/<ORIENTDB_VERSION>/libexec/bin/shutdown.sh
==> OrientDB installed, server's root user password is 'changeme'
==> Please, follow the instruction on the link below to reset it
==> http://orientdb.com/docs/2.2/Server-Security.html#restoring-the-servers-user-root
...
The installation process setups a default server's root user password that must be changed.
The orientdb-server-config.xml
file is installed in /usr/local/Cellar/orientdb/<ORIENTDB_VERSION>/libexec/config/
.
Open the file and remove the "root" user entry.
Remove the tag
/usr/local/Cellar/orientdb/<ORIENTDB_VERSION>/libexec/bin/server.sh
The script asks for a new password for the database's root user.
Other resources
To learn more about how to install OrientDB on specific environment please follow the guide below:
- Install on Linux Ubuntu
- Install on JBoss AS
- Install on GlassFish
- Install on Ubuntu 12.04 VPS (DigitalOcean)
- Install as service on Unix, Linux and MacOSX
- Install as service on Windows
Install as a Service on Windows
OrientDB is a Java server application. As most server applications, they have to perform several tasks, before being able to shut down the Virtual Machine process, hence they need a portable way to be notified of the imminent Virtual Machine shutdown. At the moment, the only way to properly shut down an OrientDB server instance (not embedded) is to execute the shutdown.bat (or shutdown.sh) script shipped with the OrientDB distribution, but it's up to the user to take care of this. This implies that the server instance isn't stopped correctly, when the computer on which it is deployed, is shut down without executing the above script.
IMPORTANT Note that you need to start the OrientDB server once manually via server.bat
in %ORIENTDB_HOME%\bin once, before starting the service.
Apache Commons Daemon
Apache Commons Daemon is a set of applications and API enabling Java server application to run as native non interactive server applications under Unix and Windows. In Unix, server applications running in the background are called daemons and are controlled by the operating system with a set of specified signals. Under Windows, such programs are called services and are controlled by appropriate calls to specific functions defined in the application binary. Although the ways of dealing with running daemons or services are different, in both cases the operating system can notify a server application of its imminent shutdown, and the underlying application has the ability to perform certain tasks, before its process of execution is destroyed. Wrapping OrientDB as a Unix daemon or as a Windows service enables the management of this server application lifecycle through the mechanisms provided natively by both Unix and Windows operating systems.
Installation
This tutorial is focused on Windows, so you have to download procrun. Procrun is a set of applications, which allow Windows users to wrap (mostly) Java applications (e.g. Tomcat) as a Windows service. The service can be set to automatically start, when the machine boots and will continue to run with no user logged onto the machine.
- Point you browser to the Apache Commons Daemon download page.
- Click on Browse native binaries download area...: you will see the index commons/daemon/binaries/ (even if the title in the page reports Index of dist/commons).
- Click on windows. Now you can see the index of commons/daemon/binaries/windows.
- Click on commons-daemon-x.x.x-bin-windows.zip. The download starts.
- Unzip the file in a directory of your choice. The content of the archive is depicted below:
commons-daemon-1.0.7-bin-windows
|
\---amd64
|
\---prunsrv.exe
|
\---ia64
|
\---prunsrv.exe
|
\---LICENCE.txt
|
\---NOTICE.txt
|
\---prunmgr.exe
|
\---prunsrv.exe
|
\---RELEASE-NOTES.txt
prunmgr is a GUI application for monitoring and configuring Windows services wrapped with procrun. prunsrv is a service application for running applications as services. It can convert any application (not just Java applications) to run as a service. The directory amd64 contains a version of prunsrv for x86-64 machines while the directory ia64 contains a version of prunsrv for Itanium 64 machines.
Once you downloaded the applications, you have to put them in a folder under the OrientDB installation folder.
- Go to the OrientDB folder, in the following referred as %ORIENTDB_HOME%
- Create a new directory and name it service
- Copy there the appropriate versions of prunsrv according to the architecture of your machine and prunmgr.
Configuration
In this section, we will show how to wrap OrientDB as a Windows Service. In order to wrap OrientDB as a service, you have to execute a short script that uses the prunsrv application to configure a Windows Service.
Before defining the Windows Service, you have to rename prunsrv and prunmgr according to the name of the service. Both applications require the name of the service to manage and monitor as parameter but you can avoid it by naming them with the name of the service. In this case, rename them respectively OrientDBGraph and OrientDBGraphw as OrientDBGraph is the name of the service that you are going to configure with the script below. If you want to use a difference service name, you have to rename both application respectively myservicename and myservicenamew (for example, if you are wrapping OrientDB and the name of the service is OrientDB, you could rename prunsrv as OrientDB and prunmgr as OrientDBw). After that, create the file %ORIENTDB_HOME%\service\installService.bat with the content depicted below:
:: OrientDB Windows Service Installation
@echo off
rem Remove surrounding quotes from the first parameter
set str=%~1
rem Check JVM DLL location parameter
if "%str%" == "" goto missingJVM
set JVM_DLL=%str%
rem Remove surrounding quotes from the second parameter
set str=%~2
rem Check OrientDB Home location parameter
if "%str%" == "" goto missingOrientDBHome
set ORIENTDB_HOME=%str%
set CONFIG_FILE=%ORIENTDB_HOME%/config/orientdb-server-config.xml
set LOG_FILE=%ORIENTDB_HOME%/config/orientdb-server-log.properties
set LOG_CONSOLE_LEVEL=info
set LOG_FILE_LEVEL=fine
set WWW_PATH=%ORIENTDB_HOME%/www
set ORIENTDB_ENCODING=UTF8
set ORIENTDB_SETTINGS=-Dprofiler.enabled=true -Dcache.level1.enabled=false -Dcache.level2.strategy=1
set JAVA_OPTS_SCRIPT=-XX:+HeapDumpOnOutOfMemoryError
rem Install service
OrientDBGraph.exe //IS --DisplayName="OrientDB GraphEd" ^
--Description="OrientDB Graph Edition, aka GraphEd, contains OrientDB server integrated with the latest release of the TinkerPop Open Source technology stack supporting property graph data model." ^
--StartClass=com.orientechnologies.orient.server.OServerMain --StopClass=com.orientechnologies.orient.server.OServerShutdownMain ^
--Classpath="%ORIENTDB_HOME%\lib\*" --JvmOptions=-Dfile.encoding=%ORIENTDB_ENCODING%;-Djava.util.logging.config.file="%LOG_FILE%";-Dorientdb.config.file="%CONFIG_FILE%";-Dorientdb.www.path="%WWW_PATH%";-Dlog.console.level=%LOG_CONSOLE_LEVEL%;-Dlog.file.level=%LOG_FILE_LEVEL%;-Dorientdb.build.number="@BUILD@";-DORIENTDB_HOME="%ORIENTDB_HOME%" ^
--StartMode=jvm --StartPath="%ORIENTDB_HOME%\bin" --StopMode=jvm --StopPath="%ORIENTDB_HOME%\bin" --Jvm="%JVM_DLL%" --LogPath="%ORIENTDB_HOME%\log" --Startup=auto
EXIT /B
:missingJVM
echo Insert the JVM DLL location
goto printUsage
:missingOrientDBHome
echo Insert the OrientDB Home
goto printUsage
:printUsage
echo usage:
echo installService JVM_DLL_location OrientDB_Home
EXIT /B
The script requires two input parameters:
- The location of jvm.dll, for example C:\Program Files\Java\jdk1.6.0_26\jre\bin\server\jvm.dll
- The location of the OrientDB installation folder, for example D:\orientdb-graphed-1.0rc5
The service is actually installed when executing OrientDBGraph.exe (originally prunsrv) with the appropriate set of command line arguments and parameters. The command line argument //IS states that the execution of that application will result in a service installation. Below there is the table with the command line parameters used in the above script.
Parameter name | Description | Source |
---|---|---|
--DisplayName | The name displayed in the Windows Services Management Console | Custom |
--Description | The description displayed in the Windows Services Management Console | Custom |
--StartClass | Class that contains the startup method (= the method to be called to start the application). The default method to be called is the main method | The class invoked in the */bin/server.bat* script |
--StopClass | Class that will be used when receiving a Stop service signal. The default method to be called is the main method | The class invoked in the */bin/shutdown.bat* script |
--Classpath | Set the Java classpath | The value of the -cp parameter specified in the _%ORIENTDB_HOME%\bin\server.bat_ script |
--JvmOptions | List of options to be passed to the JVM separated using either # or ; characters | The list of options in the form of -D or -X specified in the _%ORIENTDB_HOME%\bin\server.bat_ script and the definition of the ORIENTDB_HOME system property |
--StartMode | Specify how to start the process. In this case, it will start Java in-process and not as a separate image | Based on Apache Tomcat configuration |
--StartPath | Working path for the StartClass | _%ORIENTDB_HOME%\bin_ |
--StopMode | The same as --StartMode | Based on Apache Tomcat configuration |
--StopPath | Working path for the StopClass | _%ORIENTDB_HOME%\bin_ |
--Jvm | Which *jvm.dll* to use: the default one or the one located in the specified full path | The first input parameter of this script. Ensure that you insert the location of the Java HotSpot Server VM as a full path. We will use the server version for both start and stop. |
--LogPath | Path used by prunsrv for logging | The default location of the Apache Commons Daemon log |
--Startup | States if the service should start at machine start up or manually | auto |
In order to install the service:
- Open the Windows command shell
- Go to %ORIENTDB_HOME%\service, for example typing in the shell
> cd D:\orientdb-graphed-1.0rc5\service
- Execute the installService.bat specifying the jvm.dll location and the OrientDB Home as full paths, for example typing in the shell
> installService.bat "C:\Program Files\Java\jdk1.6.0_26\jre\bin\server\jvm.dll" D:\orientdb-graphed-1.0rc5
- Open the Windows Services Management Console - from the taskbar, click on Start, Control Panel, Administrative Tools and then Service - and check the existance of a service with the same name specified as value of the
--DisplayName
parameter (in this case OrientDB GraphEd 1.0rc5). You can also use %ORIENTDB_HOME%\service\OrientDBGraphw.exe to manage and monitor the OrientDBGraph service.
Example (in the service
directroy) start and stop the service:
:start
OrientDBGraph.exe //ES
:stop
OrientDBGraph.exe //SS
Note that you need to start the OrientDB server once manually via server.bat
in %ORIENTDB_HOME%\bin once, before starting the service.
Other resources
To learn more about how to install OrientDB on specific environment please follow the guide below:
- Install on Linux Ubuntu
- Install on JBoss AS
- Install on GlassFish
- Install on Ubuntu 12.04 VPS (DigitalOcean)
- Install as service on Unix, Linux and MacOSX
- Install as service on Windows
Installing in a Docker Container
If you have Docker installed in your computer, this is the easiest way to run OrientDB. From the command line type:
$ docker run -d --name orientdb -p 2424:2424 -p 2480:2480
-e ORIENTDB_ROOT_PASSWORD=root orientdb:latest
Where instead of "root", type the root's password you want to use.
Building the image on your own
Dockerfiles are available on a dedicated repository. The repository has a folder for each maintained version of OrientDB. Dockerfiles are approved by Docker's team.. This allows to build images on your own or even customize them for your special purpose.
- Clone this project to a local folder:
git clone https://github.com/orientechnologies/orientdb-docker.git
- Build the image for 2.2.x:
cd 2.2 docker build -t <YOUR_DOCKER_HUB_USER>/orientdb:2.2.11 .
- Push it to your Docker Hub repository (it will ask for your login credentials):
docker push <YOUR_DOCKER_HUB_USER>/orientdb:2.2.11
Configuration
OrientDB can be configured in several ways.
Getting settings
From Java
To retrieve the default or current value of a setting you can use the methods getDefValue()
and getValue()
of the java enumerator OGlobalConfiguration
.
Example:
- get the default value of the setting
USE_WAL
:OGlobalConfiguration.USE_WAL.getDefValue();
- get the current value of the setting
USE_WAL
:OGlobalConfiguration.USE_WAL.getValue());
From the Console
To retrieve the full list of current settings from the Console you can use the CONFIG
command:
orientdb> CONFIG
LOCAL SERVER CONFIGURATION
+----+----------------------------------------------+--------------+
|# |NAME |VALUE |
+----+----------------------------------------------+--------------+
|0 |environment.dumpCfgAtStartup |false |
|1 |environment.concurrent |true |
|2 |environment.lockManager.concurrency.level |32 |
|3 |environment.allowJVMShutdown |true |
|4 |script.pool.maxSize |20 |
...
|219 |client.channel.minPool |1 |
|220 |storage.keepOpen |true |
|221 |cache.local.enabled |true |
+----+----------------------------------------------+--------------+
To retrieve the value of a specific setting, you can use the CONFIG GET
command.
Example:
orientdb> CONFIG GET tx.log.fileType
Remote configuration: tx.log.fileType = classic
Changing settings
By command line
You can pass settings via command line when the JVM is launched. This is typically stored inside server.sh
(or server.bat
on Windows):
java -Dcache.size=10000 -Dstorage.keepOpen=true ...
By server configuration
Put in the <properties>
section of the file orientdb-server-config.xml (or orientdb-dserver-config.xml) the entries to configure.
Example:
...
<properties>
<entry name="cache.size" value="10000" />
<entry name="storage.keepOpen" value="true" />
</properties>
...
At run-time
Some settings can be changed at run-time.
From Java
To change at run-time a setting from java you can use the setValue()
method of the java enumerator OGlobalConfiguration
.
Example:
OGlobalConfiguration.MVRBTREE_NODE_PAGE_SIZE.setValue(2048);
From the Console
To change at run-time a setting from the Console you can use the CONFIG SET
command.
Example:
orientdb> CONFIG SET tx.autoRetry 5
Dump the configuration
To dump the OrientDB configuration you can set a parameter at JVM launch:
java -Denvironment.dumpCfgAtStartup=true ...
Or via API at any time:
OGlobalConfiguration.dumpConfiguration(System.out);
Parameters
To know more look at the Java enumeration: OGlobalConfiguration.java. NOTE: The documentation below is auto-generated by parsing the OGlobalConfiguration.java class. Please do not edit the parameters here, but rather edit the documentation in the java class.
Environment
environment.dumpCfgAtStartup
Dumps the configuration during application startup.
Setting name...: environment.dumpCfgAtStartup
Default value..: false
Set at run-time: false
environment.concurrent
Specifies if running in multi-thread environment. Setting this to false turns off the internal lock management.
Setting name...: environment.concurrent
Default value..: true
Set at run-time: false
environment.lockManager.concurrency.level
Concurrency level of lock manager.
Setting name...: environment.lockManager.concurrency.level
Default value..: 64
Set at run-time: false
environment.allowJVMShutdown
Allows the shutdown of the JVM, if needed/requested.
Setting name...: environment.allowJVMShutdown
Default value..: true
Set at run-time: true
Script
script.pool.maxSize
Maximum number of instances in the pool of script engines.
Setting name...: script.pool.maxSize
Default value..: 20
Set at run-time: false
Memory
memory.useUnsafe
Indicates whether Unsafe will be used, if it is present.
Setting name...: memory.useUnsafe
Default value..: true
Set at run-time: false
memory.chunk.size
Size of single memory chunk (in bytes) which will be preallocated by OrientDB.
Setting name...: memory.chunk.size
Default value..: 2147483647
Set at run-time: false
memory.directMemory.safeMode
Indicates whether to perform a range check before each direct memory update. It is true by default, but usually it can be safely set to false. It should only be to true after dramatic changes have been made in the storage structures.
Setting name...: memory.directMemory.safeMode
Default value..: true
Set at run-time: false
memory.directMemory.trackMode
Activates the direct memory pool leak detector. This detector causes a large overhead and should be used for debugging purposes only. It's also a good idea to pass the -Djava.util.logging.manager=com.orientechnologies.common.log.OLogManager$DebugLogManager switch to the JVM, if you use this mode, this will enable the logging from JVM shutdown hooks..
Setting name...: memory.directMemory.trackMode
Default value..: false
Set at run-time: false
memory.directMemory.onlyAlignedMemoryAccess
Some architectures do not allow unaligned memory access or may suffer from speed degradation. For such platforms, this flag should be set to true.
Setting name...: memory.directMemory.onlyAlignedMemoryAccess
Default value..: true
Set at run-time: false
Jvm
jvm.gc.delayForOptimize
Minimal amount of time (in seconds), since the last System.gc(), when called after tree optimization.
Setting name...: jvm.gc.delayForOptimize
Default value..: 600
Set at run-time: false
Storage
storage.openFiles.limit
Limit of amount of files which may be open simultaneously.
Setting name...: storage.openFiles.limit
Default value..: 512
Set at run-time: false
storage.componentsLock.cache
Amount of cached locks is used for component lock to avoid constant creation of new lock instances.
Setting name...: storage.componentsLock.cache
Default value..: 10000
Set at run-time: false
storage.diskCache.pinnedPages
Maximum amount of pinned pages which may be contained in cache, if this percent is reached next pages will be left in unpinned state. You can not set value more than 50.
Setting name...: storage.diskCache.pinnedPages
Default value..: 20
Set at run-time: false
storage.diskCache.bufferSize
Size of disk buffer in megabytes, disk size may be changed at runtime, but if does not enough to contain all pinned pages exception will be thrown.
Setting name...: storage.diskCache.bufferSize
Default value..: 4096
Set at run-time: true
storage.diskCache.writeCachePart
Percentage of disk cache, which is used as write cache.
Setting name...: storage.diskCache.writeCachePart
Default value..: 15
Set at run-time: false
storage.diskCache.writeCachePageTTL
Max time until a page will be flushed from write cache (in seconds).
Setting name...: storage.diskCache.writeCachePageTTL
Default value..: 86400
Set at run-time: false
storage.diskCache.writeCachePageFlushInterval
Interval between flushing of pages from write cache (in ms).
Setting name...: storage.diskCache.writeCachePageFlushInterval
Default value..: 25
Set at run-time: false
storage.diskCache.writeCacheFlushInactivityInterval
Interval between 2 writes to the disk cache, if writes are done with an interval more than provided, all files will be fsynced before the next write, which allows a data restore after a server crash (in ms).
Setting name...: storage.diskCache.writeCacheFlushInactivityInterval
Default value..: 60000
Set at run-time: false
storage.diskCache.writeCacheFlushLockTimeout
Maximum amount of time the write cache will wait before a page flushes (in ms, -1 to disable).
Setting name...: storage.diskCache.writeCacheFlushLockTimeout
Default value..: -1
Set at run-time: false
storage.diskCache.diskFreeSpaceCheckInterval
The interval (in seconds), after which the storage periodically checks whether the amount of free disk space is enough to work in write mode.
Setting name...: storage.diskCache.diskFreeSpaceCheckInterval
Default value..: 5
Set at run-time: false
storage.diskCache.diskFreeSpaceCheckIntervalInPages
The interval (how many new pages should be added before free space will be checked), after which the storage periodically checks whether the amount of free disk space is enough to work in write mode.
Setting name...: storage.diskCache.diskFreeSpaceCheckIntervalInPages
Default value..: 2048
Set at run-time: false
storage.diskCache.keepState
Keep disk cache state between moment when storage is closed and moment when it is opened again. true by default.
Setting name...: storage.diskCache.keepState
Default value..: true
Set at run-time: false
storage.configuration.syncOnUpdate
Indicates a force sync should be performed for each update on the storage configuration.
Setting name...: storage.configuration.syncOnUpdate
Default value..: true
Set at run-time: false
storage.compressionMethod
Record compression method used in storage Possible values : gzip, nothing, snappy, snappy-native. Default is 'nothing' that means no compression.
Setting name...: storage.compressionMethod
Default value..: nothing
Set at run-time: false
storage.encryptionMethod
Record encryption method used in storage Possible values : 'aes' and 'des'. Default is 'nothing' for no encryption.
Setting name...: storage.encryptionMethod
Default value..: nothing
Set at run-time: false
storage.encryptionKey
Contains the storage encryption key. This setting is hidden.
Setting name...: storage.encryptionKey
Default value..: null
Set at run-time: false
storage.makeFullCheckpointAfterCreate
Indicates whether a full checkpoint should be performed, if storage was created.
Setting name...: storage.makeFullCheckpointAfterCreate
Default value..: false
Set at run-time: false
storage.makeFullCheckpointAfterOpen
Indicates whether a full checkpoint should be performed, if storage was opened. It is needed so fuzzy checkpoints can work properly.
Setting name...: storage.makeFullCheckpointAfterOpen
Default value..: true
Set at run-time: false
storage.makeFullCheckpointAfterClusterCreate
Indicates whether a full checkpoint should be performed, if storage was opened.
Setting name...: storage.makeFullCheckpointAfterClusterCreate
Default value..: true
Set at run-time: false
storage.trackChangedRecordsInWAL
If this flag is set metadata which contains rids of changed records is added at the end of each atomic operation.
Setting name...: storage.trackChangedRecordsInWAL
Default value..: false
Set at run-time: false
storage.useWAL
Whether WAL should be used in paginated storage.
Setting name...: storage.useWAL
Default value..: true
Set at run-time: false
storage.wal.syncOnPageFlush
Indicates whether a force sync should be performed during WAL page flush.
Setting name...: storage.wal.syncOnPageFlush
Default value..: true
Set at run-time: false
storage.wal.cacheSize
Maximum size of WAL cache (in amount of WAL pages, each page is 64k) If set to 0, caching will be disabled.
Setting name...: storage.wal.cacheSize
Default value..: 3000
Set at run-time: false
storage.wal.fileAutoCloseInterval
Interval in seconds after which WAL file will be closed if there is no any IO operations on this file (in seconds), default value is 10.
Setting name...: storage.wal.fileAutoCloseInterval
Default value..: 10
Set at run-time: false
storage.wal.maxSegmentSize
Maximum size of single WAL segment (in megabytes).
Setting name...: storage.wal.maxSegmentSize
Default value..: 128
Set at run-time: false
storage.wal.maxSize
Maximum size of WAL on disk (in megabytes).
Setting name...: storage.wal.maxSize
Default value..: 4096
Set at run-time: false
storage.wal.commitTimeout
Maximum interval between WAL commits (in ms.).
Setting name...: storage.wal.commitTimeout
Default value..: 1000
Set at run-time: false
storage.wal.shutdownTimeout
Maximum wait interval between events, when the background flush threadreceives a shutdown command and when the background flush will be stopped (in ms.).
Setting name...: storage.wal.shutdownTimeout
Default value..: 10000
Set at run-time: false
storage.wal.fuzzyCheckpointInterval
Interval between fuzzy checkpoints (in seconds).
Setting name...: storage.wal.fuzzyCheckpointInterval
Default value..: 300
Set at run-time: false
storage.wal.reportAfterOperationsDuringRestore
Amount of processed log operations, after which status of data restore procedure will be printed (0 or a negative value, disables the logging).
Setting name...: storage.wal.reportAfterOperationsDuringRestore
Default value..: 10000
Set at run-time: false
storage.wal.restore.batchSize
Amount of WAL records, which are read at once in a single batch during a restore procedure.
Setting name...: storage.wal.restore.batchSize
Default value..: 1000
Set at run-time: false
storage.wal.readCacheSize
Size of WAL read cache in amount of pages.
Setting name...: storage.wal.readCacheSize
Default value..: 1000
Set at run-time: false
storage.wal.fuzzyCheckpointShutdownWait
The amount of time the DB should wait until it shuts down (in seconds).
Setting name...: storage.wal.fuzzyCheckpointShutdownWait
Default value..: 600
Set at run-time: false
storage.wal.fullCheckpointShutdownTimeout
The amount of time the DB will wait, until a checkpoint is finished, during a DB shutdown (in seconds).
Setting name...: storage.wal.fullCheckpointShutdownTimeout
Default value..: 600
Set at run-time: false
storage.wal.path
Path to the WAL file on the disk. By default, it is placed in the DB directory, but it is highly recommended to use a separate disk to store log operations.
Setting name...: storage.wal.path
Default value..: null
Set at run-time: false
storage.diskCache.pageSize
Size of page of disk buffer (in kilobytes). !!! NEVER CHANGE THIS VALUE !!!.
Setting name...: storage.diskCache.pageSize
Default value..: 64
Set at run-time: false
storage.diskCache.diskFreeSpaceLimit
Minimum amount of space on disk, which, when exceeded, will cause the database to switch to read-only mode (in megabytes).
Setting name...: storage.diskCache.diskFreeSpaceLimit
Default value..: 256
Set at run-time: false
storage.lowestFreeListBound
The least amount of free space (in kb) in a page, which is tracked in paginated storage.
Setting name...: storage.lowestFreeListBound
Default value..: 16
Set at run-time: false
storage.lockTimeout
Maximum amount of time (in ms) to lock the storage.
Setting name...: storage.lockTimeout
Default value..: 0
Set at run-time: false
storage.record.lockTimeout
Maximum of time (in ms) to lock a shared record.
Setting name...: storage.record.lockTimeout
Default value..: 2000
Set at run-time: false
storage.useTombstones
When a record is deleted, the space in the cluster will not be freed, but rather tombstoned.
Setting name...: storage.useTombstones
Default value..: false
Set at run-time: false
storage.cluster.usecrc32
Indicates whether crc32 should be used for each record to check record integrity.
Setting name...: storage.cluster.usecrc32
Default value..: false
Set at run-time: false
storage.keepOpen
Deprecated.
Setting name...: storage.keepOpen
Default value..: true
Set at run-time: false
Record
record.downsizing.enabled
On updates, if the record size is lower than before, this reduces the space taken accordingly. If enabled this could increase defragmentation, but it reduces the used disk space.
Setting name...: record.downsizing.enabled
Default value..: true
Set at run-time: false
Object
object.saveOnlyDirty
Object Database only! It saves objects bound to dirty records.
Setting name...: object.saveOnlyDirty
Default value..: false
Set at run-time: true
Db
db.pool.min
Default database pool minimum size.
Setting name...: db.pool.min
Default value..: 1
Set at run-time: false
db.pool.max
Default database pool maximum size.
Setting name...: db.pool.max
Default value..: 100
Set at run-time: false
db.pool.idleTimeout
Timeout for checking for free databases in the pool.
Setting name...: db.pool.idleTimeout
Default value..: 0
Set at run-time: false
db.pool.idleCheckDelay
Delay time on checking for idle databases.
Setting name...: db.pool.idleCheckDelay
Default value..: 0
Set at run-time: false
db.mvcc.throwfast
Use fast-thrown exceptions for MVCC OConcurrentModificationExceptions. No context information will be available. Set to true, when these exceptions are thrown, but the details are not necessary.
Setting name...: db.mvcc.throwfast
Default value..: false
Set at run-time: true
db.validation
Enables or disables validation of records.
Setting name...: db.validation
Default value..: true
Set at run-time: true
db.document.serializer
The default record serializer used by the document database.
Setting name...: db.document.serializer
Default value..: ORecordSerializerBinary
Set at run-time: false
db.makeFullCheckpointOnIndexChange
When index metadata is changed, a full checkpoint is performed.
Setting name...: db.makeFullCheckpointOnIndexChange
Default value..: true
Set at run-time: true
db.makeFullCheckpointOnSchemaChange
When index schema is changed, a full checkpoint is performed.
Setting name...: db.makeFullCheckpointOnSchemaChange
Default value..: true
Set at run-time: true
db.mvcc
Deprecated, MVCC cannot be disabled anymore.
Setting name...: db.mvcc
Default value..: true
Set at run-time: false
db.use.distributedVersion
Deprecated, distributed version is not used anymore.
Setting name...: db.use.distributedVersion
Default value..: false
Set at run-time: false
NonTX
nonTX.recordUpdate.synch
Executes a sync against the file-system for every record operation. This slows down record updates, but guarantees reliability on unreliable drives.
Setting name...: nonTX.recordUpdate.synch
Default value..: false
Set at run-time: false
nonTX.clusters.sync.immediately
List of clusters to sync immediately after update (separated by commas). Can be useful for a manual index.
Setting name...: nonTX.clusters.sync.immediately
Default value..: manindex
Set at run-time: false
Tx
tx.trackAtomicOperations
This setting is used only for debug purposes. It creates a stack trace of methods, when an atomic operation is started.
Setting name...: tx.trackAtomicOperations
Default value..: false
Set at run-time: false
tx.commit.synch
Synchronizes the storage after transaction commit.
Setting name...: tx.commit.synch
Default value..: false
Set at run-time: false
tx.autoRetry
Maximum number of automatic retry if some resource has been locked in the middle of the transaction (Timeout exception).
Setting name...: tx.autoRetry
Default value..: 1
Set at run-time: false
tx.log.fileType
File type to handle transaction logs: mmap or classic.
Setting name...: tx.log.fileType
Default value..: classic
Set at run-time: false
tx.log.synch
Executes a synch against the file-system at every log entry. This slows down transactions but guarantee transaction reliability on unreliable drives.
Setting name...: tx.log.synch
Default value..: false
Set at run-time: false
tx.useLog
Transactions use log file to store temporary data to be rolled back in case of crash.
Setting name...: tx.useLog
Default value..: true
Set at run-time: false
Index
index.embeddedToSbtreeBonsaiThreshold
Amount of values, after which the index implementation will use an sbtree as a values container. Set to -1, to disable and force using an sbtree.
Setting name...: index.embeddedToSbtreeBonsaiThreshold
Default value..: 40
Set at run-time: true
index.sbtreeBonsaiToEmbeddedThreshold
Amount of values, after which index implementation will use an embedded values container (disabled by default).
Setting name...: index.sbtreeBonsaiToEmbeddedThreshold
Default value..: -1
Set at run-time: true
index.auto.synchronousAutoRebuild
Synchronous execution of auto rebuilding of indexes, in case of a DB crash.
Setting name...: index.auto.synchronousAutoRebuild
Default value..: true
Set at run-time: false
index.auto.lazyUpdates
Configure the TreeMaps for automatic indexes, as buffered or not. -1 means buffered until tx.commit() or db.close() are called.
Setting name...: index.auto.lazyUpdates
Default value..: 10000
Set at run-time: false
index.flushAfterCreate
Flush storage buffer after index creation.
Setting name...: index.flushAfterCreate
Default value..: true
Set at run-time: false
index.manual.lazyUpdates
Configure the TreeMaps for manual indexes as buffered or not. -1 means buffered until tx.commit() or db.close() are called.
Setting name...: index.manual.lazyUpdates
Default value..: 1
Set at run-time: false
index.durableInNonTxMode
Indicates whether index implementation for plocal storage will be durable in non-Tx mode (true by default).
Setting name...: index.durableInNonTxMode
Default value..: true
Set at run-time: false
index.ignoreNullValuesDefault
Controls whether null values will be ignored by default by newly created indexes or not (false by default).
Setting name...: index.ignoreNullValuesDefault
Default value..: false
Set at run-time: false
index.txMode
Indicates the index durability level in TX mode. Can be ROLLBACK_ONLY or FULL (ROLLBACK_ONLY by default).
Setting name...: index.txMode
Default value..: FULL
Set at run-time: false
index.cursor.prefetchSize
Default prefetch size of index cursor.
Setting name...: index.cursor.prefetchSize
Default value..: 500000
Set at run-time: false
index.auto.rebuildAfterNotSoftClose
Auto rebuild all automatic indexes after upon database open when wasn't closed properly.
Setting name...: index.auto.rebuildAfterNotSoftClose
Default value..: true
Set at run-time: false
HashTable
hashTable.slitBucketsBuffer.length
Length of buffer (in pages), where buckets that were split, but not flushed to the disk, are kept. This buffer is used to minimize random IO overhead.
Setting name...: hashTable.slitBucketsBuffer.length
Default value..: 1500
Set at run-time: false
Sbtree
sbtree.maxDepth
Maximum depth of sbtree, which will be traversed during key look up until it will be treated as broken (64 by default).
Setting name...: sbtree.maxDepth
Default value..: 64
Set at run-time: false
sbtree.maxKeySize
Maximum size of a key, which can be put in the SBTree in bytes (10240 by default).
Setting name...: sbtree.maxKeySize
Default value..: 10240
Set at run-time: false
sbtree.maxEmbeddedValueSize
Maximum size of value which can be put in an SBTree without creation link to a standalone page in bytes (40960 by default).
Setting name...: sbtree.maxEmbeddedValueSize
Default value..: 40960
Set at run-time: false
Sbtreebonsai
sbtreebonsai.bucketSize
Size of bucket in OSBTreeBonsai (in kB). Contract: bucketSize < storagePageSize, storagePageSize % bucketSize == 0.
Setting name...: sbtreebonsai.bucketSize
Default value..: 2
Set at run-time: false
sbtreebonsai.linkBagCache.size
Amount of LINKBAG collections to be cached, to avoid constant reloading of data.
Setting name...: sbtreebonsai.linkBagCache.size
Default value..: 100000
Set at run-time: false
sbtreebonsai.linkBagCache.evictionSize
The number of cached LINKBAG collections, which will be removed, when the cache limit is reached.
Setting name...: sbtreebonsai.linkBagCache.evictionSize
Default value..: 1000
Set at run-time: false
sbtreebonsai.freeSpaceReuseTrigger
How much free space should be in an sbtreebonsai file, before it will be reused during the next allocation.
Setting name...: sbtreebonsai.freeSpaceReuseTrigger
Default value..: 0.5
Set at run-time: false
RidBag
ridBag.embeddedDefaultSize
Size of embedded RidBag array, when created (empty).
Setting name...: ridBag.embeddedDefaultSize
Default value..: 4
Set at run-time: false
ridBag.embeddedToSbtreeBonsaiThreshold
Amount of values after which a LINKBAG implementation will use sbtree as values container. Set to -1 to always use an sbtree.
Setting name...: ridBag.embeddedToSbtreeBonsaiThreshold
Default value..: 40
Set at run-time: true
ridBag.sbtreeBonsaiToEmbeddedToThreshold
Amount of values, after which a LINKBAG implementation will use an embedded values container (disabled by default).
Setting name...: ridBag.sbtreeBonsaiToEmbeddedToThreshold
Default value..: -1
Set at run-time: true
Collections
collections.preferSBTreeSet
This configuration setting is experimental.
Setting name...: collections.preferSBTreeSet
Default value..: false
Set at run-time: false
File
file.trackFileClose
Log all the cases when files are closed. This is needed only for internal debugging purposes.
Setting name...: file.trackFileClose
Default value..: false
Set at run-time: false
file.lock
Locks files when used. Default is true.
Setting name...: file.lock
Default value..: true
Set at run-time: false
file.deleteDelay
Delay time (in ms) to wait for another attempt to delete a locked file.
Setting name...: file.deleteDelay
Default value..: 10
Set at run-time: false
file.deleteRetry
Number of retries to delete a locked file.
Setting name...: file.deleteRetry
Default value..: 50
Set at run-time: false
Security
security.userPasswordSaltIterations
Number of iterations to generate the salt or user password. Changing this setting does not affect stored passwords.
Setting name...: security.userPasswordSaltIterations
Default value..: 65536
Set at run-time: false
security.userPasswordSaltCacheSize
Cache size of hashed salt passwords. The cache works as LRU. Use 0 to disable the cache.
Setting name...: security.userPasswordSaltCacheSize
Default value..: 500
Set at run-time: false
security.userPasswordDefaultAlgorithm
Default encryption algorithm used for passwords hashing.
Setting name...: security.userPasswordDefaultAlgorithm
Default value..: PBKDF2WithHmacSHA256
Set at run-time: false
security.createDefaultUsers
Indicates whether default database users should be created.
Setting name...: security.createDefaultUsers
Default value..: true
Set at run-time: false
Network
network.maxConcurrentSessions
Maximum number of concurrent sessions.
Setting name...: network.maxConcurrentSessions
Default value..: 1000
Set at run-time: true
network.socketBufferSize
TCP/IP Socket buffer size.
Setting name...: network.socketBufferSize
Default value..: 32768
Set at run-time: true
network.lockTimeout
Timeout (in ms) to acquire a lock against a channel.
Setting name...: network.lockTimeout
Default value..: 15000
Set at run-time: true
network.socketTimeout
TCP/IP Socket timeout (in ms).
Setting name...: network.socketTimeout
Default value..: 15000
Set at run-time: true
network.requestTimeout
Request completion timeout (in ms).
Setting name...: network.requestTimeout
Default value..: 3600000
Set at run-time: true
network.retry
Number of attempts to connect to the server on failure.
Setting name...: network.retry
Default value..: 5
Set at run-time: true
network.retryDelay
The time (in ms) the client must wait, before reconnecting to the server on failure.
Setting name...: network.retryDelay
Default value..: 500
Set at run-time: true
network.binary.loadBalancing.enabled
Asks for DNS TXT record, to determine if load balancing is supported.
Setting name...: network.binary.loadBalancing.enabled
Default value..: false
Set at run-time: true
network.binary.loadBalancing.timeout
Maximum time (in ms) to wait for the answer from DNS about the TXT record for load balancing.
Setting name...: network.binary.loadBalancing.timeout
Default value..: 2000
Set at run-time: true
network.binary.maxLength
TCP/IP max content length (in KB) of BINARY requests.
Setting name...: network.binary.maxLength
Default value..: 16384
Set at run-time: true
network.binary.readResponse.maxTimes
Maximum attempts, until a response can be read. Otherwise, the response will be dropped from the channel.
Setting name...: network.binary.readResponse.maxTimes
Default value..: 20
Set at run-time: true
network.binary.debug
Debug mode: print all data incoming on the binary channel.
Setting name...: network.binary.debug
Default value..: false
Set at run-time: true
network.http.installDefaultCommands
Installs the default HTTP commands.
Setting name...: network.http.installDefaultCommands
Default value..: true
Set at run-time: true
network.http.serverInfo
Server info to send in HTTP responses. Change the default if you want to hide it is a OrientDB Server.
Setting name...: network.http.serverInfo
Default value..: OrientDB Server v.2.2.12-SNAPSHOT
Set at run-time: true
network.http.maxLength
TCP/IP max content length (in bytes) for HTTP requests.
Setting name...: network.http.maxLength
Default value..: 1000000
Set at run-time: true
network.http.streaming
Enable Http chunked streaming for json responses.
Setting name...: network.http.streaming
Default value..: true
Set at run-time: false
network.http.charset
Http response charset.
Setting name...: network.http.charset
Default value..: utf-8
Set at run-time: true
network.http.jsonResponseError
Http response error in json.
Setting name...: network.http.jsonResponseError
Default value..: true
Set at run-time: true
network.http.jsonp
Enable the usage of JSONP, if requested by the client. The parameter name to use is 'callback'.
Setting name...: network.http.jsonp
Default value..: false
Set at run-time: true
network.http.sessionExpireTimeout
Timeout, after which an http session is considered to have expired (in seconds).
Setting name...: network.http.sessionExpireTimeout
Default value..: 300
Set at run-time: false
network.http.useToken
Enable Token based sessions for http.
Setting name...: network.http.useToken
Default value..: false
Set at run-time: false
network.token.secretKey
Network token secret key.
Setting name...: network.token.secretKey
Default value..:
Set at run-time: false
network.token.encryptionAlgorithm
Network token algorithm.
Setting name...: network.token.encryptionAlgorithm
Default value..: HmacSHA256
Set at run-time: false
network.token.expireTimeout
Timeout, after which a binary session is considered to have expired (in minutes).
Setting name...: network.token.expireTimeout
Default value..: 60
Set at run-time: false
Profiler
profiler.enabled
Enables the recording of statistics and counters.
Setting name...: profiler.enabled
Default value..: false
Set at run-time: true
profiler.config
Configures the profiler as
Setting name...: profiler.config
Default value..: null
Set at run-time: true
profiler.autoDump.interval
Dumps the profiler values at regular intervals (in seconds).
Setting name...: profiler.autoDump.interval
Default value..: 0
Set at run-time: true
profiler.maxValues
Maximum values to store. Values are managed in a LRU.
Setting name...: profiler.maxValues
Default value..: 200
Set at run-time: false
Sequence
sequence.maxRetry
Maximum number of retries between attempt to change a sequence in concurrent mode.
Setting name...: sequence.maxRetry
Default value..: 100
Set at run-time: false
sequence.retryDelay
Maximum number of ms to wait between concurrent modification exceptions. The value is computed as random between 1 and this number.
Setting name...: sequence.retryDelay
Default value..: 200
Set at run-time: false
StorageProfiler
storageProfiler.intervalBetweenSnapshots
Interval between snapshots of profiler state in milliseconds.
Setting name...: storageProfiler.intervalBetweenSnapshots
Default value..: 100
Set at run-time: false
storageProfiler.cleanUpInterval
Interval between time series in milliseconds.
Setting name...: storageProfiler.cleanUpInterval
Default value..: 5000
Set at run-time: false
Log
log.console.level
Console logging level.
Setting name...: log.console.level
Default value..: info
Set at run-time: true
log.file.level
File logging level.
Setting name...: log.file.level
Default value..: info
Set at run-time: true
log.console.ansi
ANSI Console support. 'auto' means automatic check if it is supported, 'true' to force using ANSI, 'false' to avoid using ANSI.
Setting name...: log.console.ansi
Default value..: auto
Set at run-time: false
Class
class.minimumClusters
Minimum clusters to create when a new class is created. 0 means Automatic.
Setting name...: class.minimumClusters
Default value..: 0
Set at run-time: false
Cache
cache.local.impl
Local Record cache implementation.
Setting name...: cache.local.impl
Default value..: com.orientechnologies.orient.core.cache.ORecordCacheWeakRefs
Set at run-time: false
cache.local.enabled
Deprecated, Level1 cache cannot be disabled anymore.
Setting name...: cache.local.enabled
Default value..: true
Set at run-time: false
Command
command.timeout
Default timeout for commands (in ms).
Setting name...: command.timeout
Default value..: 0
Set at run-time: true
command.cache.enabled
Enable command cache.
Setting name...: command.cache.enabled
Default value..: false
Set at run-time: false
command.cache.evictStrategy
Command cache strategy between: [INVALIDATE_ALL,PER_CLUSTER].
Setting name...: command.cache.evictStrategy
Default value..: PER_CLUSTER
Set at run-time: false
command.cache.minExecutionTime
Minimum execution time to consider caching the result set.
Setting name...: command.cache.minExecutionTime
Default value..: 10
Set at run-time: false
command.cache.maxResultsetSize
Maximum resultset time to consider caching result set.
Setting name...: command.cache.maxResultsetSize
Default value..: 500
Set at run-time: false
Query
query.parallelAuto
Auto enable parallel query, if requirements are met.
Setting name...: query.parallelAuto
Default value..: false
Set at run-time: false
query.parallelMinimumRecords
Minimum number of records to activate parallel query automatically.
Setting name...: query.parallelMinimumRecords
Default value..: 300000
Set at run-time: false
query.parallelResultQueueSize
Size of the queue that holds results on parallel execution. The queue is blocking, so in case the queue is full, the query threads will be in a wait state.
Setting name...: query.parallelResultQueueSize
Default value..: 20000
Set at run-time: false
query.scanPrefetchPages
Pages to prefetch during scan. Setting this value higher makes scans faster, because it reduces the number of I/O operations, though it consumes more memory. (Use 0 to disable).
Setting name...: query.scanPrefetchPages
Default value..: 20
Set at run-time: false
query.scanBatchSize
Scan clusters in blocks of records. This setting reduces the lock time on the cluster during scans. A high value mean a faster execution, but also a lower concurrency level. Set to 0 to disable batch scanning. Disabling batch scanning is suggested for read-only databases only.
Setting name...: query.scanBatchSize
Default value..: 1000
Set at run-time: false
query.scanThresholdTip
If the total number of records scanned in a query exceeds this setting, then a warning is given. (Use 0 to disable).
Setting name...: query.scanThresholdTip
Default value..: 50000
Set at run-time: false
query.limitThresholdTip
If the total number of returned records exceeds this value, then a warning is given. (Use 0 to disable).
Setting name...: query.limitThresholdTip
Default value..: 10000
Set at run-time: false
query.live.support
Enable/Disable the support of live query. (Use false to disable).
Setting name...: query.live.support
Default value..: true
Set at run-time: false
Statement
statement.cacheSize
Number of parsed SQL statements kept in cache.
Setting name...: statement.cacheSize
Default value..: 100
Set at run-time: false
Sql
sql.graphConsistencyMode
Consistency mode for graphs. It can be 'tx' (default), 'notx_sync_repair' and 'notx_async_repair'. 'tx' uses transactions to maintain consistency. Instead both 'notx_sync_repair' and 'notx_async_repair' do not use transactions, and the consistency, in case of JVM crash, is guaranteed by a database repair operation that run at startup. With 'notx_sync_repair' the repair is synchronous, so the database comes online after the repair is ended, while with 'notx_async_repair' the repair is a background process.
Setting name...: sql.graphConsistencyMode
Default value..: tx
Set at run-time: false
Client
client.channel.maxPool
Maximum size of pool of network channels between client and server. A channel is a TCP/IP connection.
Setting name...: client.channel.maxPool
Default value..: 100
Set at run-time: false
client.connectionPool.waitTimeout
Maximum time, where the client should wait for a connection from the pool, when all connections busy.
Setting name...: client.connectionPool.waitTimeout
Default value..: 5000
Set at run-time: true
client.channel.dbReleaseWaitTimeout
Delay (in ms), after which a data modification command will be resent, if the DB was frozen.
Setting name...: client.channel.dbReleaseWaitTimeout
Default value..: 10000
Set at run-time: true
client.ssl.enabled
Use SSL for client connections.
Setting name...: client.ssl.enabled
Default value..: false
Set at run-time: false
client.ssl.keyStore
Use SSL for client connections.
Setting name...: client.ssl.keyStore
Default value..: null
Set at run-time: false
client.ssl.keyStorePass
Use SSL for client connections.
Setting name...: client.ssl.keyStorePass
Default value..: null
Set at run-time: false
client.ssl.trustStore
Use SSL for client connections.
Setting name...: client.ssl.trustStore
Default value..: null
Set at run-time: false
client.ssl.trustStorePass
Use SSL for client connections.
Setting name...: client.ssl.trustStorePass
Default value..: null
Set at run-time: false
client.krb5.config
Location of the Kerberos configuration file.
Setting name...: client.krb5.config
Default value..: null
Set at run-time: false
client.krb5.ccname
Location of the Kerberos client ticketcache.
Setting name...: client.krb5.ccname
Default value..: null
Set at run-time: false
client.krb5.ktname
Location of the Kerberos client keytab.
Setting name...: client.krb5.ktname
Default value..: null
Set at run-time: false
client.credentialinterceptor
The name of the CredentialInterceptor class.
Setting name...: client.credentialinterceptor
Default value..: null
Set at run-time: false
client.session.tokenBased
Request a token based session to the server.
Setting name...: client.session.tokenBased
Default value..: true
Set at run-time: false
client.channel.minPool
Minimum pool size.
Setting name...: client.channel.minPool
Default value..: 1
Set at run-time: false
Server
server.openAllDatabasesAtStartup
If true, the server opens all the available databases at startup. Available since 2.2.
Setting name...: server.openAllDatabasesAtStartup
Default value..: false
Set at run-time: false
server.channel.cleanDelay
Time in ms of delay to check pending closed connections.
Setting name...: server.channel.cleanDelay
Default value..: 5000
Set at run-time: false
server.cache.staticFile
Cache static resources upon loading.
Setting name...: server.cache.staticFile
Default value..: false
Set at run-time: false
server.log.dumpClientExceptionLevel
Logs client exceptions. Use any level supported by Java java.util.logging.Level class: OFF, FINE, CONFIG, INFO, WARNING, SEVERE.
Setting name...: server.log.dumpClientExceptionLevel
Default value..: FINE
Set at run-time: false
server.log.dumpClientExceptionFullStackTrace
Dumps the full stack trace of the exception sent to the client.
Setting name...: server.log.dumpClientExceptionFullStackTrace
Default value..: false
Set at run-time: true
server.security.file
Location of the OrientDB security.json configuration file.
Setting name...: server.security.file
Default value..: null
Set at run-time: false
Distributed
distributed.crudTaskTimeout
Maximum timeout (in ms) to wait for CRUD remote tasks.
Setting name...: distributed.crudTaskTimeout
Default value..: 3000
Set at run-time: true
distributed.commandTaskTimeout
Maximum timeout (in ms) to wait for command distributed tasks.
Setting name...: distributed.commandTaskTimeout
Default value..: 120000
Set at run-time: true
distributed.commandQuickTaskTimeout
Maximum timeout (in ms) to wait for quick command distributed tasks.
Setting name...: distributed.commandQuickTaskTimeout
Default value..: 5000
Set at run-time: true
distributed.commandLongTaskTimeout
Maximum timeout (in ms) to wait for Long-running distributed tasks.
Setting name...: distributed.commandLongTaskTimeout
Default value..: 86400000
Set at run-time: true
distributed.deployDbTaskTimeout
Maximum timeout (in ms) to wait for database deployment.
Setting name...: distributed.deployDbTaskTimeout
Default value..: 1200000
Set at run-time: true
distributed.deployChunkTaskTimeout
Maximum timeout (in ms) to wait for database chunk deployment.
Setting name...: distributed.deployChunkTaskTimeout
Default value..: 15000
Set at run-time: true
distributed.deployDbTaskCompression
Compression level (between 0 and 9) to use in backup for database deployment.
Setting name...: distributed.deployDbTaskCompression
Default value..: 7
Set at run-time: true
distributed.asynchQueueSize
Queue size to handle distributed asynchronous operations. The bigger is the queue, the more operation are buffered, but also more memory it's consumed. 0 = dynamic allocation, which means up to 2^31-1 entries.
Setting name...: distributed.asynchQueueSize
Default value..: 0
Set at run-time: false
distributed.asynchResponsesTimeout
Maximum timeout (in ms) to collect all the asynchronous responses from replication. After this time the operation is rolled back (through an UNDO).
Setting name...: distributed.asynchResponsesTimeout
Default value..: 15000
Set at run-time: false
distributed.purgeResponsesTimerDelay
Maximum timeout (in ms) to collect all the asynchronous responses from replication. This is the delay the purge thread uses to check asynchronous requests in timeout.
Setting name...: distributed.purgeResponsesTimerDelay
Default value..: 15000
Set at run-time: false
distributed.conflictResolverRepairerChain
Chain of conflict resolver implementation to use.
Setting name...: distributed.conflictResolverRepairerChain
Default value..: majority,content,version
Set at run-time: false
distributed.conflictResolverRepairerCheckEvery
Time (in ms) when the conflict resolver auto-repairer checks for records/cluster to repair.
Setting name...: distributed.conflictResolverRepairerCheckEvery
Default value..: 5000
Set at run-time: true
distributed.conflictResolverRepairerBatch
Number of record to repair in batch.
Setting name...: distributed.conflictResolverRepairerBatch
Default value..: 100
Set at run-time: true
distributed.txAliveTimeout
Maximum timeout (in ms) a distributed transaction can be alive. This timeout is to rollback pending transactions after a while.
Setting name...: distributed.txAliveTimeout
Default value..: 30000
Set at run-time: true
distributed.requestChannels
Number of network channels used to send requests.
Setting name...: distributed.requestChannels
Default value..: 1
Set at run-time: false
distributed.responseChannels
Number of network channels used to send responses.
Setting name...: distributed.responseChannels
Default value..: 1
Set at run-time: false
distributed.heartbeatTimeout
Maximum time in ms to wait for the heartbeat. If the server does not respond in time, it is put offline.
Setting name...: distributed.heartbeatTimeout
Default value..: 10000
Set at run-time: false
distributed.checkHealthCanOfflineServer
In case a server does not respond to the heartbeat message, it is set offline.
Setting name...: distributed.checkHealthCanOfflineServer
Default value..: false
Set at run-time: false
distributed.checkHealthEvery
Time in ms to check the cluster health. Set to 0 to disable it.
Setting name...: distributed.checkHealthEvery
Default value..: 10000
Set at run-time: false
distributed.autoRemoveOfflineServers
This is the amount of time (in ms) the server has to be OFFLINE, before it is automatically removed from the distributed configuration. -1 = never, 0 = immediately, >0 the actual time to wait.
Setting name...: distributed.autoRemoveOfflineServers
Default value..: -1
Set at run-time: true
distributed.publishNodeStatusEvery
Time in ms to publish the node status on distributed map. Set to 0 to disable such refresh of node configuration.
Setting name...: distributed.publishNodeStatusEvery
Default value..: 5000
Set at run-time: true
distributed.localQueueSize
Size of the intra-thread queue for distributed messages.
Setting name...: distributed.localQueueSize
Default value..: 10000
Set at run-time: false
distributed.dbWorkerThreads
Number of parallel worker threads per database that process distributed messages.
Setting name...: distributed.dbWorkerThreads
Default value..: 8
Set at run-time: false
distributed.queueMaxSize
Maximum queue size to mark a node as stalled. If the number of messages in queue are more than this values, the node is restarted with a remote command (0 = no maximum, which means up to 2^31-1 entries).
Setting name...: distributed.queueMaxSize
Default value..: 10000
Set at run-time: false
distributed.backupDirectory
Directory where the copy of an existent database is saved, before it is downloaded from the cluster. Leave it empty to avoid the backup..
Setting name...: distributed.backupDirectory
Default value..: ../backup/databases
Set at run-time: false
distributed.concurrentTxMaxAutoRetry
Maximum attempts the transaction coordinator should execute a transaction automatically, if records are locked. (Minimum is 1 = no attempts).
Setting name...: distributed.concurrentTxMaxAutoRetry
Default value..: 10
Set at run-time: true
distributed.atomicLockTimeout
Timeout (in ms) to acquire a distributed lock on a record. (0=infinite).
Setting name...: distributed.atomicLockTimeout
Default value..: 300
Set at run-time: true
distributed.concurrentTxAutoRetryDelay
Delay (in ms) between attempts on executing a distributed transaction, which had failed because of locked records. (0=no delay).
Setting name...: distributed.concurrentTxAutoRetryDelay
Default value..: 100
Set at run-time: true
distributed.queueTimeout
Maximum timeout (in ms) to wait for the response in replication.
Setting name...: distributed.queueTimeout
Default value..: 500000
Set at run-time: true
Jna
jna.disable.system.library
This property disables using JNA, should it be installed on your system. (Default true) To use JNA bundled with database.
Setting name...: jna.disable.system.library
Default value..: true
Set at run-time: false
Oauth2
oauth2.secretkey
Http OAuth2 secret key.
Setting name...: oauth2.secretkey
Default value..:
Set at run-time: false
Lazyset
lazyset.workOnStream
Deprecated, now BINARY serialization is used in place of CSV.
Setting name...: lazyset.workOnStream
Default value..: true
Set at run-time: false
Mvrbtree
mvrbtree.timeout
Deprecated, MVRBTREE IS NOT USED ANYMORE IN FAVOR OF SBTREE AND HASHINDEX.
Setting name...: mvrbtree.timeout
Default value..: 0
Set at run-time: false
mvrbtree.nodePageSize
Deprecated, MVRBTREE IS NOT USED ANYMORE IN FAVOR OF SBTREE AND HASHINDEX.
Setting name...: mvrbtree.nodePageSize
Default value..: 256
Set at run-time: false
mvrbtree.loadFactor
Deprecated, MVRBTREE IS NOT USED ANYMORE IN FAVOR OF SBTREE AND HASHINDEX.
Setting name...: mvrbtree.loadFactor
Default value..: 0.7
Set at run-time: false
mvrbtree.optimizeThreshold
Deprecated, MVRBTREE IS NOT USED ANYMORE IN FAVOR OF SBTREE AND HASHINDEX.
Setting name...: mvrbtree.optimizeThreshold
Default value..: 100000
Set at run-time: false
mvrbtree.entryPoints
Deprecated, MVRBTREE IS NOT USED ANYMORE IN FAVOR OF SBTREE AND HASHINDEX.
Setting name...: mvrbtree.entryPoints
Default value..: 64
Set at run-time: false
mvrbtree.optimizeEntryPointsFactor
Deprecated, MVRBTREE IS NOT USED ANYMORE IN FAVOR OF SBTREE AND HASHINDEX.
Setting name...: mvrbtree.optimizeEntryPointsFactor
Default value..: 1.0
Set at run-time: false
mvrbtree.entryKeysInMemory
Deprecated, MVRBTREE IS NOT USED ANYMORE IN FAVOR OF SBTREE AND HASHINDEX.
Setting name...: mvrbtree.entryKeysInMemory
Default value..: false
Set at run-time: false
mvrbtree.entryValuesInMemory
Deprecated, MVRBTREE IS NOT USED ANYMORE IN FAVOR OF SBTREE AND HASHINDEX.
Setting name...: mvrbtree.entryValuesInMemory
Default value..: false
Set at run-time: false
mvrbtree.ridBinaryThreshold
Deprecated, MVRBTREE IS NOT USED ANYMORE IN FAVOR OF SBTREE AND HASHINDEX.
Setting name...: mvrbtree.ridBinaryThreshold
Default value..: -1
Set at run-time: false
mvrbtree.ridNodePageSize
Deprecated, MVRBTREE IS NOT USED ANYMORE IN FAVOR OF SBTREE AND HASHINDEX.
Setting name...: mvrbtree.ridNodePageSize
Default value..: 64
Set at run-time: false
mvrbtree.ridNodeSaveMemory
Deprecated, MVRBTREE IS NOT USED ANYMORE IN FAVOR OF SBTREE AND HASHINDEX.
Setting name...: mvrbtree.ridNodeSaveMemory
Default value..: false
Set at run-time: false
NOTE: On 64-bit systems you have not the limitation of 32-bit systems with memory.
Logging
Logging is configured in a separate file, look at Logging for more information.
Storage configuration
OrientDB allows modifications to the storage configuration. Even though this will be supported with high level commands, for now it's pretty "internal" using Java API.
To get the storage configuration for the current database:
OStorageConfiguration cfg = db.getStorage().getConfiguration();
Look at OStorageConfiguration
to discover all the properties you can change. To change the configuration of a cluster get it by ID;
OStoragePhysicalClusterConfigurationLocal clusterCfg = (OStoragePhysicalClusterConfigurationLocal) cfg.clusters.get(3);
To change the default settings for new clusters get the file template object. In this example we change the initial file size from the default 500Kb down to 10Kb:
OStorageSegmentConfiguration defaultCfg = (OStorageSegmentConfiguration) cfg.fileTemplate;
defaultCfg.fileStartSize = "10Kb";
After changes call OStorageConfiguration.update()
:
cfg.update();
Distributed Configuration
For information about the Distributed Configuration, please refer to the section Distributed Configuration.
Upgrading
OrientDB uses the Semantic Versioning System (http://semver.org), where the version numbers follow this format MAJOR.MINOR.PATCH, Here are the meanings of the increments:
- MAJOR version entails incompatible API changes,
- MINOR version entails functionality in a backward-compatible manner
- PATCH version entails backward-compatible bug fixes.
So between PATCH versions, the compatibility is assured (example 2.1.0 -> 2.1.8). Between MINOR and MAJOR versions, you may need to export and re-import the database. To find out if your upgrade must be done over exporting and importing the database, see below in the column "Database":
Compatibility Matrix
FROM | TO | Guide | Blueprints | Database | Binary Protocol | HTTP Protocol |
---|---|---|---|---|---|---|
3.1.x | 3.2.x | Docs | v 3.3.0 | Automatic | 38 | 10 |
3.0.x | 3.1.x | No need to upgrade the application layer (import/export only) | v 3.3.0 | Automatic | 38 | 10 |
2.2.x | 3.0.x | Docs | v 3.3.0 | Automatic | 37 | 10 |
2.1.x | 2.2.x | Docs | Final v2.6.0 | Automatic | 34 | 10 |
2.0.x | 2.1.x | Docs | Final v2.6.0 | Automatic | 30 | 10 |
Instructions
There are two possible scenarios:
- upgrade to a new hotfix release (eg. from 3.0.10 to 3.0.11)
- upgrade to a new major/minor release (eg. from 3.0.31 to 3.1.0)
For major/minor release upgrades, please refer to the docs linked in the above table.
For hotfix upgrades, the easiest way to upgrade OrientDB is to plan ahead for future upgrades from the beginning.
The recommended strategy is to store databases separately from the OrientDB installation, often on a separate data volume.
As an example, you might have a data volume, called /data
, with a databases
directory under that where all of your database directories will be stored.
/data/databases:
MyDB
WebAppDB
Upgrading with no existing databases
If you're just starting with OrientDB or just creating new databases, from the OrientDB installation directory, you can remove the databases
directory and then create a symbolic link to the /data/databases
directory previously created.
Make sure OrientDB is not running.
On a Linux system, you could call rm -rf databases
to remove the databases directory and recursively remove all sub-directories.
DO NOT issue that command if you have any live databases that have not been moved!
Once the databases
directory is removed, create a symbolic link to the /data/databases
directory.
On a Linux system, you could call ln -s /data/databases ./databases
.
On a Windows system, you can use mklink /d .\databases d:\data\databases
, assuming your data volume resides as the d:
drive in Windows. The /d
switch creates a directory symbolic link. You can use /j
instead to create a directory junction.
You should now be able to start OrientDB and create new databases that will reside under /data/databases
.
Upgrading Symbolic-Linked Databases
If you used a similar symbolic link scheme as suggested in the prior section, upgrading OrientDB is very easy. Simply follow the same instructions to remove the databases
directory from the NEW OrientDB installation directory and then create a symbolic link to /data/databases
.
Upgrading Existing Databases
If you've been running OrientDB in the standard way, with the databases
directory stored directly under the OrientDB installation directory, then you have two options when upgrading to a newer version of OrientDB.
- You can move your database directories to the
databases
directory under the new installation. - You can move your database directories to an external location and then set-up a symbolic link from the new installation.
Moving Databases to a New Installation
Make sure OrientDB is not running. From the old OrientDB installation location, move each database directory under databases
to the databases
directory in the new OrientDB installation.
Moving Databases to an External Location
Make sure OrientDB is not running. Using the earlier example of /data/databases
, from the old OrientDB installation location, move each database directory under databases
to the /data/databases
location.
Now, follow the instructions under New Databases to create a symbolic link from within the new OrientDB installation to the /data/databases
directory.
External Storage Advantages
If you store your databases separately from the OrientDB installation, not only will you be able to upgrade more easily in the future, but you may even be able to improve performance by using a data volume that is mounted on a disk that's faster than the volume where the main installation resides.
Also, many cloud-based providers support creating snapshots of data volumes, which can be a useful way of backing up all of your databases.
Backward Compatibility
When upgrading OrientDB to a more recent version, you should always follow the migration instructions (see upgrade guide ).
This said, in OrientDB development process we always try to support binary compatibility between two subsequent minor releases.
For example, lets suppose that we have following releases 1.5, 1.5.1, 1.6.1, 1.6.2, 1.7, 1.7.1 then binary compatibility at least between 1.6.1, 1.6.2, 1.7, 1.7.1 releases will be supported.
If we have releases 1.5, 1.5.1, 1.6.1, 1.6.2, 1.7, 1.7.1, 2.0 then binary compatibility will be supported at least between releases 1.7, 1.7.1, 2.0.
Binary compatibility feature is implemented using following algorithm:
- When storage is opened, version of binary format which is used when storage is created is read from storage configuration.
- Factory of objects are used to present disk based data structures for current binary format is created.
Only features and database components which existed at the moment when current binary format was latest one will be used. It means that you cannot use all database features available in latest release if you use storage which was created using old binary format version. It also means that bugs which are fixed in new versions may be (but may be not) reproducible on storage created using old binary format.
To update binary format storage to latest one you should export database in JSON format and import it back.
Using either console commands export database and import database or Java API look at com.orientechnologies.orient.core.db.tool.ODatabaseImport
, com.orientechnologies.orient.core.db.tool.ODatabaseExport
classes and com.orientechnologies.orient.test.database.auto.DbImportExportTest
test.
- Current binary format version can be read from
com.orientechnologies.orient.core.db.record.OCurrentStorageComponentsFactory#binaryFormatVersion
property. - Instance of
OCurrentStorageComponentsFactory
class can be retrieved by call ofcom.orientechnologies.orient.core.storage.OStorage#getComponentsFactory
method. - Latest binary format version can be read from here
com.orientechnologies.orient.core.config.OStorageConfiguration#CURRENT_BINARY_FORMAT_VERSION
.
Please note that binary compatibility is supported since 1.7-rc2 version for plocal storage (as exception you can read database created in 1.5.1 version by 1.7-rc2 version).
Return to Upgrade.
Upgrading a Distributed Environment
Logging
OrientDB handles logs using the Java Logging Framework, which is bundled with the JVM. The specific format it uses derives from the OLogFormatter
class, which defaults to:
<date> <level> <message> [<requester>]
<date>
Shows the date of the log entry, using the date formatYYYY-MM-DD HH:MM:SS:SSS
.<level>
Shows the log level.<message>
Shows the log message.<class>
Shows the Java class that made the entry, (optional).
The supported levels are those contained in the JRE class java.util.logging.Level
. From highest to lowest:
SEVERE
WARNING
INFO
CONFIG
FINE
FINER
FINEST
By default, OrientDB installs two loggers:
console
: Logs to the shell or command-prompt that starts the application or the server. You can modify it by setting thelog.console.level
variable.file
: Logs to the log file. You can modify it by setting thelog.file.level
variable.
Configuration File
You can configure logging strategies and policies by creating a configuration file that follows the
Java Logging Messages configuration syntax. For example, consider the following from the orientdb-server-log.properties
file:
# Specify the handlers to create in the root logger
# (all loggers are children of the root logger)
# The following creates two handlers
handlers = java.util.logging.ConsoleHandler, java.util.logging.FileHandler
# Set the default logging level for the root logger
.level = ALL
# Set the default logging level for new ConsoleHandler instances
java.util.logging.ConsoleHandler.level = INFO
# Set the default formatter for new ConsoleHandler instances
java.util.logging.ConsoleHandler.formatter = com.orientechnologies.common.log.OLogFormatter
# Set the default logging level for new FileHandler instances
java.util.logging.FileHandler.level = INFO
# Naming style for the output file
java.util.logging.FileHandler.pattern=../log/orient-server.log
# Set the default formatter for new FileHandler instances
java.util.logging.FileHandler.formatter = com.orientechnologies.common.log.OLogFormatter
# Limiting size of output file in bytes:
java.util.logging.FileHandler.limit=10000000
# Number of output files to cycle through, by appending an
# integer to the base file name:
java.util.logging.FileHandler.count=10
When the log properties file is ready, you need to tell the JVM to use t, by setting java.util.logging.config.file
system property.
$ java -Djava.util.logging.config.file=mylog.properties
Setting the Log Level
To change the log level without modifying the logging configuration, set the log.console.level
and log.file.level
system variables. These system variables are accessible both at startup and at runtime.
Configuring Log Level at Startup
You can configure log level at startup through both the orientdb-server-config.xml
configuration file and by modifying the JVM before you start the server:
Using the Configuration File
To configure log level from the configuration file, update the following elements in the <properties>
section:
<properties>
<entry value="info" name="log.console.level" />
<entry value="fine" name="log.file.level" />
...
</properties>
Using the JVM
To configure log level from the JVM before starting the server, run the java
command to configure the log.console.level
and log.file.level
variables:
$ java -Dlog.console.level=INFO -Dlog.file.level=FINE
Configuring Log Level at Runtime
You can configure log level at runtime through both the Java API and by executing an HTTP POST
against the remote server.
Using Java Code
Through the Java API, you can set the system variables for logging at startup through the System.setProperty()
method. For instance,
public void main(String[] args){
System.setProperty("log.console.level", "FINE");
...
}
Using HTTP POST
Through the HTTP requests, you can update the logging system variables by executing a POST
against the URL: /server/log.<type>/<level>
.
<type>
Defines the log type:console
orfile
.<level>
Defines the log level.
Examples
The examples below use cURL to execute the HTTP POST
commands against the OrientDB server. It uses the server root
user and password.
-
Enable the finest tracing level to the console:
$
curl -u root:root -X POST http://localhost:2480/server/log.console/FINEST
-
Enable the finest tracing level to file:
$
curl -u root:root -X POST http://localhost:2480/server/log.file/FINEST
Change logging on the client
On the client is the same as for the server, but you should rather configure the file config/orientdb-client-log.properties
and add this at your client's JVM:
$ java -Djava.util.logging.config.file=config/orientdb-client-log.properties
Install Log Formatter
OrientDB Server uses its own log formatter. In order to enable the same for your application, you need to include the following line:
OLogManager.installCustomFormatter();
The Server automatically installs the log formatter. To disable it, use orientdb.installCustomFormatter
.
$ java -Dorientdb.installCustomFormatter=false
Debugging Logger
Java Logging Framework runtime has a known problem with logging from shutdown hooks, sometimes log entries may be lost. OrientDB uses shutdown hooks to properly shutdown its internals, so it also may be affected by this problem, something may go wrong silently on shutdown. To workaround this problem OrientDB provides a special logger – the debugging logger. To activate it provide following command line argument to your JVM:
-Djava.util.logging.manager=com.orientechnologies.common.log.OLogManager$DebugLogManager
Use this logger for debugging and troubleshooting purposes only, since it may interfere with your production logging configuration.
Make sure
$DebugLogManager
part is not interpreted as a shell variable substitution. To avoid the substitution apply escaping specific to your shell environment.
Event Scheduler
OrientDB has a Scheduler of events you can use to fire your events on regular basis. To manage events you can use both SQL and Java API. The scheduler gets the popular CRON expression syntax. The scheduled event, when fired, executes a Database
Resources
- CRON expressions on Wikipedia
- CRON expression maker is an online resource to create CRON expressions
Schedule an event
Via SQL
In order to schedule a new event via SQL, all you need is to create a new record in the OSchedule
class. Example on scheduling the event "myEvent" that calls the function "myFunction" every second:
INSERT INTO oschedule
SET name = 'myEvent',
function = (SELECT FROM ofunction WHERE name = 'myFunction'),
rule = \"0/1 * * * * ?\"
Via Java API
db.getMetadata().getScheduler().scheduleEvent(
new OScheduledEventBuilder().setName("myEvent").setRule("0/1 * * * * ?")
.setFunction(func).build());
Update an event
Via SQL
To update the scheduling of an event, update the record with the new rule
. Example:
UPDATE oschedule SET rule = "0/2 * * * * ?" WHERE name = 'myEvent'
Via Java API
To update an event, remove it and reschedule it.
db.getMetadata().getScheduler().removeEvent("myEvent");
db.getMetadata().getScheduler().scheduleEvent(
new OScheduledEventBuilder().setName("myEvent").setRule("0/2 * * * * ?")
.setFunction(func).build());
Remove an event
Via SQL
To cancel a scheduled event, just delete the record. Example:
DELETE oschedule WHERE name = 'myEvent'
Via Java API
db.getMetadata().getScheduler().removeEvent("myEvent");
Tutorial
In this tutorial we want to purge all the records older than 1 year.
1) Create a Function
First, create a SQL function that delete the records. To have the date of 1y ago you can use the expression sysdate() - 31536000000
, where 31536000000 represents the number of milliseconds in a year. You can this via SQL or Java API.
Via SQL
CREATE FUNCTION purgeHistoricalRecords
"DELETE FROM Logs WHERE date < ( sysdate() - 31536000000 )"
LANGUAGE SQL
Via Java API
OFunction func = db.getMetadata().getFunctionLibrary().createFunction("purgeHistoricalRecords");
func.setLanguage("SQL");
func.setCode("DELETE FROM Logs WHERE date < ( sysdate() - 31536000000 )");
func.save();
return func;
2) Schedule the event
The second step is scheduling the event. The CRON expression for "every midnight" is 00 00 * * * ?
. You can this via SQL or Java API.
Via SQL
INSERT INTO oschedule
SET name = 'purgeHistoricalRecordsAtMidnight',
function = (SELECT FROM ofunction WHERE name = 'purgeHistoricalRecords'),
rule = \"00 00 * * * ?\"
Via Java API
db.command(new OCommandSQL(
"INSERT INTO oschedule SET name = 'purgeHistoricalRecordsAtMidnight', function = ?, rule = \"00 00 * * * ?\""))
.execute(func.getId());
Functions
A Function is an executable unit of code that can take parameters and return a result. Using Functions you can perform Functional programming where logic and data are all together in a central place. Functions are similar to the Stored Procedures of RDBMS.
NOTE: This guide refers to the last available release of OrientDB. For past revisions look at Compatibility.
OrientDB Functions features:
- are persistent
- can be written in SQL or Javascript (Ruby, Scala, Java and other languages are coming)
- can be executed via SQL, Java, REST and Studio
- can call each other
- supports recursion
- have automatic mapping of parameters by position and name
- plugins can inject new objects to being used by functions
The OrientDB SQL dialect supports many functions written in the native language. To get better performance, you can write your own native functions in the Java language and register them to the engine.
For more information, see Custom Functions in Java.
Compatibility
1.5.0 and before
OrientDB binds the following variables:
db
, that is the current document database instancegdb
, that is the current graph database instance
Creating Functions
In addition to the functions provided in OrientDB by default, you can also create your own functions.
Using OrientDB Studio
You can do so through the OrientDB Studio. To create a sample function, complete the following steps:
-
In Studio, open a database and navigate to the Functions panel.
-
Enter
sum
as a name, then add two parameters:a
andb
. -
Give the function the following code in the text area:
return parseInt(a) + parseInt(b);
-
Click Save.
This adds the function sum
to the database. In Studio, it appears on the left-side with the available functions.
On the bottom of the page, there are two empty boxes in which you can test the function. Write 3
in one and 5
in the other, then click Execute to see the results.
Why using
parseInt()
instead ofa + b
? Because HTTP protocol passes parameters as strings.
Saved Functions
OrientDB saves functions in the database, using the OFunction
class. This class has the following properties:
Property | Description |
---|---|
name | Defines the name of the function. |
code | Defines the code the function executes. |
parameters | Defines an optional EMBEDDEDLIST of strings, containing the parameter names, if any. |
idempotent | Defines whether the function is idempotent, that is if it changes the database. Read-only functions are idempotent. This is needed to avoid calling non-idempotent functions using the HTTP GET method. |
Given that OrientDB uses one record per function, the MVCC mechanism is used to protect against concurrent record updates.
Using Recursive Calls
Functions can call on other functions in order to process the results. They can also call themselves in a recursive call.
For instance, create a new function with the name factorial
. Give it the parameter num
, with the following code:
if (num === 0)
return 1;
else
return num * factorial( num - 1 );
When you call factorial()
, it calls itself in order to determine the factorial of the given num
value, until it reaches the result, which in the case of the test is 3648800.0
.
Using Functions
Call Functions via SQL
All the database functions are automatically registered in the SQL Engine.
Examples
-
Calling the sum function from SQL with static parameters:
SELECT SUM(3,4)
-
Calling the sum function from SQL with parameters taken from database:
SELECT SUM(salary,bonus) AS total FROM Employee
In this case the function SUM is invoked for each record in the Employee class by passing the value of the fields
salary
andbonus
.
Functions using the Java API
Using OrientDB functions from the Java API is relatively straightforward:
- Get the reference to the Function Manager
- Get the function you want to use.
- Execute the function.
Examples
-
Passing the parameters by position:
ODatabaseDocumentTx db = new ODatabaseDocumentTx("local:/tmp/db"); db.open("admin", "admin"); OFunction sum = db.getMetadata().getFunctionLibrary().getFunction("sum"); Number result = sum.execute(3, 5);
-
Using the Blueprints Graph API:
OFunction sum = graph.getRawGraph().getMetadata().getFunctionLibrary().getFunction("sum");
-
Execute the function by passing the parameters by name:
Map<String,Object> params = new HashMap<String,Object>();
params.put("a", 3);
params.put("b", 5);
Number result = sum.execute(params);
Functions using the HTTP REST API
OrientDB exposes functions as a REST service, through which it can receive parameters. Parameters are passed by position in the URL. Beginning with OrientDB version 2.1, you can also pass parameters in the request payload as JSON. In which case, the mapping is not positional, but by name.
Examples
-
Execute the
sum
function, passing3
and5
as parameters in the URL:http://localhost:2480/function/demo/sum/3/5
-
Execute the
sum
function, passing3
and5
in the request's payload:{"a": 3, "b": 5}
Each example returns an HTTP 202 OK with an envelope containing the result of the calculation:
{"result":[{"@type":"d","@version":0,"value":2}]}
You can only call functions with the HTTP GET method if you declare it as idempotent
. You can call any functions using the HTTP POST method.
When executing functions with the HTTP POST method, encode the content and set the HTTP request header to: "Content-Type: application/json"
.
For more information, see
Function Return Values in HTTP
When calling a function through a REST service, OrientDB returns the results as JSON to the client through HTTP. There may be differences in the results, depending on the return value of function.
For instance,
-
Function that returns a number:
return 31;
Would return the result:
{"result":[{"@type":"d","@version":0,"value":31}]}
-
Function that returns a JavaScript object:
return {"a":1, "b":"foo"}
Would return the result:
{"result":[{"@type":"d","@version":0,"value":{"a":1,"b":"foo"}}]}
-
Function that returns an array:
return [1, 2, 3]
Would return the result:
{"result":[{"@type":"d","@version":0,"value":[1,2,3]}]}
-
Function that returns a query result:
return db.query("SELECT FROM OUser")
Would return the result:
{ "result": [ { "@type": "d", "@rid": "#6:0", "@version": 1, "@class": "OUser", "name": "admin", "password": "...", "status": "ACTIVE", "roles": [ "#4:0" ], "@fieldTypes": "roles=n" }, { "@type": "d", "@rid": "#6:1", "@version": 1, "@class": "OUser", "name": "reader", "password": "...", "status": "ACTIVE", "roles": [ "#4:1" ], "@fieldTypes": "roles=n" } ] }
Accessing the Database from a Function
When you create a function for OrientDB, it always binds the special variable orient
to allow you to use OrientDB services from within the function. The most important methods are:
Function | Description |
---|---|
orient.getDatabase() | Returns the current document database instance. |
For security reason starting from OrientDB 3.0.29, the usage of Java classes is forbidden by default, with a class filter implemented in the JS engine.
To enable the access to classes or packages in your JS code change the allowedPackages
field with comma separated packages or classes.
<handler class="com.orientechnologies.orient.server.handler.OServerSideScriptInterpreter">
<parameters>
<parameter name="enabled" value="true" />
<parameter name="allowedLanguages" value="SQL" />
<!-- Comma separated packages allowed in JS scripts eg. java.math.*, java.util.ArrayList -->
<parameter name="allowedPackages" value=""/>
</parameters>
</handler>
Executing Queries
Queries are idempotent commands. To execute a query from within a function, use the query()
method. For instance,
return orient.getDatabase().query("SELECT name FROM OUser");
Queries with External Parameters
Create a new function with the name getUserRoles
with the parameter user
. Then use this code:
return orient.getDatabase().query("SELECT roles FROM OUser WHERE name = ?", name );
Here, the function binds the name
parameter as a variable in JavaScript. You can use this variable to build your query.
SQL Commands
Execute an SQL command within the function:
var results = orient.getDatabase().command("SELECT FROM Employee WHERE company = ?", [ "Orient Technologies" ] );
The command returns an array of OElement
objects
Creating Repository Classes
Functions provide an ideal place for developing the logic your application uses to access the database. You can adopt a Domain-driven design approach, allowing the function to work as a repository or as a Data Access Object.
This provides a thin (or thick, if you prefer) layer of encapsulation which may protect you from database changes.
Furthermore, each function is published and reachable via the HTTP REST protocol, allowing the automatic creation of a RESTful service.
Examples
Below are some examples of functions to build a repository for OUser
records:
function user_getAll() {
return orient.getDatabase().query("SELECT FROM OUser");
}
function user_getByName( name ){
return orient.getDatabase().query("SELECT FROM OUser WHERE name = ?", name );
}
function user_getAdmin(){
return user_getByName("admin");
}
function user_create( name, role ){
var db = orient.getDatabase();
var role = db.query("SELECT FROM ORole WHERE name = ?", roleName);
if( role == null ){
response.send(404, "Role name not found", "text/plain", "Error: role name not found" );
} else {
db.begin();
try{
var result = db.save({ "@class" : "OUser", name : "Luca", password : "Luc4", status: "ACTIVE", roles : role});
db.commit();
return result;
}catch ( err ){
db.rollback()
response.send(500, "Error on creating new user", "text/plain", err.toString() );
}
}
}
Server-side Functions
In OrientDB, you can replace the use of Servlets with server-side functions. For more information on how to call server-side functions, see Functions using the HTTP REST API.
When the HTTP REST protocol calls server-side functions, OrientDB embeds a few additional variables:
- Request Object The HTTP request, implemented by the
OHttpRequestWrapper
class. - Response Object The HTTP request response, implemented by the
OHttpResponseWrapper
class. - Util Object The utility class with helper functions, to use inside the functions, implemented by the
OFunctionUtilWrapper
class.
Request Object
OrientDB references this object as request
. For instance,
var params = request.getParameters();
Method signature | Description | Return type |
---|---|---|
getContent() | Returns the request content. | String |
getUser() | Gets the request user name. | String |
getContentType() | Returns the request content type. | String |
getHttpVersion() | Return the request HTTP version. | String |
getHttpMethod() | Return the request HTTP method called. | String |
getIfMatch() | Return the request IF-MATCH header. | String |
isMultipart() | Returns if the request is multi-part. | boolean |
getArguments() | Returns the request arguments passed in REST form. Example: /2012/10/26 . | String[] |
getArgument(<position>) | Returns the request argument by position, or null if not found. | String |
getParameters() | Returns the request parameters. | String |
getParameter(<name> ) | Returns the request parameter by name or null if not found. | String |
hasParameters(<name>* ) | Returns the number of parameters found between those passed. | Integer |
getSessionId() | Returns the session-id. | String |
getURL() | Returns the request URL. | String |
Response Object
OrientDB references this object as response
. For instance,
var db = orient.getDatabase();
var roles = db.query("select from ORole where name = ?", roleName);
if( roles == null || roles.length == 0 ){
response.send(404, "Role name not found", "text/plain", "Error: role name not found" );
} else {
db.begin();
try{
var result = db.save({ "@class" : "OUser", name : "Luca", password : "Luc4", "roles" : roles});
db.commit();
return result;
}catch ( err ){
db.rollback();
response.send(500, "Error on creating new user", "text/plain", err.toString() );
}
}
Method signature | Description | Return type |
---|---|---|
getHeader() | Returns the response additional headers. | String |
setHeader(String header) | Sets the response additional headers to send back. To specify multiple headers use the line breaks. | Request object |
getContentType() | Returns the response content type. If null will be automatically detected. | String |
setContentType(String contentType) | Sets the response content type. If null will be automatically detected. | Request object |
getCharacterSet() | Returns the response character set used. | String |
setCharacterSet(String characterSet) | Sets the response character set. | Request object |
getHttpVersion() | Returns the HTTP version. | String |
writeStatus(int httpCode, String reason) | Sets the response status as HTTP code and reason. | Request object |
writeStatus(int httpCode, String reason) | Sets the response status as HTTP code and reason. | Request object |
writeHeaders(String contentType) | Sets the response headers using the keep-alive. | Request object |
writeHeaders(String contentType, boolean keepAlive) | Sets the response headers specifying when using the keep-alive or not. | Request object |
writeLine(String content) | Writes a line in the response. A line feed will be appended at the end of the content. | Request object |
writeContent(String content) | Writes content directly to the response. | Request object |
writeRecords(List<OIdentifiable> records) | Writes records as response. The records are serialized in JSON format. | Request object |
writeRecords( List<OIdentifiable> records, String fetchPlan) | Writes records as response specifying a fetch-plan to serialize nested records. The records are serialized in JSON format. | Request object |
writeRecord(ORecord record) | Writes a record as response. The record is serialized in JSON format. | Request object |
writeRecord(ORecord record, String fetchPlan) | Writes a record as response. The record is serialized in JSON format. | Request object |
send(int code, String reason, String contentType, Object content) | Sends the complete HTTP response in one call. | Request object |
send(int code, String reason, String contentType, Object content, String headers) | Sends the complete HTTP response in one call specifying additional headers. Keep-alive is set. | Request object |
send(int code, String reason, String contentType, Object content, String headers, boolean keepAlive) | Sends the complete HTTP response in one call specifying additional headers. | Request object |
sendStream(int code, String reason, String contentType, InputStream content, long size) | Sends the complete HTTP response in one call specifying a stream as content. | Request object |
flush() | Flushes the content to the TCP/IP socket. | Request object |
Util Object
OrientDB references this object as util
. For instance,
if( util.exists(year) ){
print("\nYes, the year was passed!");
}
Method signature | Description | Return type |
---|---|---|
exists(<variable>) | Returns trues if any of the passed variables are defined. In JS, for example, a variable is defined if it's not null and not equal to undefined . | Boolean |
Backup & Restore
OrientDB supports backup and and restore operations, like any database management system.
The BACKUP DATABASE
command executes a complete backup on the currently open database. It compresses the backup the backup using the ZIP algorithm. To restore the database from the subsequent .zip
file, you can use the RESTORE DATABASE
command.
Backups and restores are much faster than the EXPORT DATABASE
and IMPORT DATABASE
commands. You can also automate backups using the Automatic Backup server plugin. Additionally, beginning with version 2.2 of Enterprise Edition OrientDB introduces major support for incremental backups.
NOTE: OrientDB Community Edition does not support backing up remote databases. OrientDB Enterprise Edition does support this feature. For more information on how to implement this with Enterprise Edition, see Remote Backups.
Backups versus Exports
During backups, the BACKUP DATABASE
command produces a consistent copy of the database. During this process, the database locks all write operations, waiting for the backup to finish. If you need perform reads and writes on the database during backups, set up a distributed cluster of nodes. To access to the non blocking backup feature, use the Enterprise Edition.
By contrast, the EXPORT DATABASE
command doesn't lock the database, allowing concurrent writes to occur during the export process. Consequentially, the export may include changes made after you initiated the export, which may result in inconsistencies.
Using the Backup Script
Beginning in version 1.7.8, OrientDB introduces a backup.sh
script found in the $ORIENTDB_HOME/bin
directory. This script allows you to initiate backups from the system console.
Syntax
./backup.sh <db-url> <user> <password> <destination> [<type>]
<db-url>
Defines the URL for the database to backup.<user>
Defines the user to run the backup.<password>
Defines the password for the user.<destination>
Defines the path to the backup file the script creates, (use the.zip
extension).<type>
Defines the backup type. Supported types:default
Locks the database during the backup.lvm
Executes an LVM copy-on-write snapshot in the background.
NOTE Non-blocking backups require that the operating system support LVM. For more information, see
Examples
-
Backup a database opened using
plocal
:$
$ORIENTDB_HOME/bin/backup.sh plocal:../database/testdb \ admin adminpasswd \ /path/to/backup.zip
-
Perform a non-blocking LVM backup, using
plocal
:$
$ORIENTDB_HOME/bin/backup.sh plocal:../database/testdb \ admin adminpasswd \ /path/to/backup.zip \ lvm
-
Perform a backup using the OrientDB Console with the
BACKUP
command:orientdb>
CONNECT PLOCAL:../database/testdb/ admin adminpasswd
orientdb>BACKUP DATABASE /path/to/backup.zip
Backup executed in 0.52 seconds.
Restoring Databases
Once you have created your backup.zip
file, you can restore it to the database either through the OrientDB Console, using the RESTORE DATABASE
command.
orientdb> RESTORE DATABASE /backups/mydb.zip
Restore executed in 6.33 seconds
Bear in mind that OrientDB does not support merging during restores. If you need to merge the old data with new writes, instead use EXPORT DATABASE
and IMPORT DATABASE
commands, instead.
For more information, see
Incremental Backup and Restore
(Since v2.2 - Enterprise Edition only)
An incremental backup generates smaller backup files by storing only the delta between two versions of the database. This is useful when you execute a backup on a regular basis and you want to avoid having to back up the entire database each time. The easiest way to execute a backup and a restore is using Studio.
NOTE: This feature is available only in the OrientDB Enterprise Edition. If you are interested in a commercial license look at OrientDB Subscription Packages.
NOTE: Lucene Indexes are not supported yet in the incremental backup/restore process. Once the incremental restore is finished the indexes rebuild is necessary see (here)[https://github.com/orientechnologies/orientdb/issues/5958]
See also
How does it work?
Every time a backup is executed, OrientDB writes a file named last-backup.json
in the database directory. This is an example of the content:
{
"lsn": 8438432,
"startedOn": "2015-08-17 10:33:23.943",
"completedOn": "2015-08-17 10:33:45.121"
}
The most important information is the lsn
field that is the WAL LSN (Last Serial Number). Thanks to this number, OrientDB is able to understand the last change in the database, so the next incremental backup will be done starting from last lsn
+ 1.
Executing an Incremental Backup
Incremental Backup via Console
Backup Database console command accepts -incremental
as an optional parameter to execute an incremental backup. In this case the new backup is executed from the last backup (file last-backup.json
is read if present). If this is the first incremental backup, a full backup is executed. Example:
orientdb> connect plocal:/databases/mydb admin admin
orientdb {db=Whisky}> backup database /tmp/backup -incremental
The incremental backup setting also allows you to specify an LSN version to start with. Example:
orientdb> connect plocal:/databases/mydb admin admin
orientdb {db=Whisky}> backup database /tmp/backup -incremental=93222
Incremental Backup via Java API
You can perform an incremental backup via the Java API too.
NOTE The remote
protocol is supported, but the specified path is relative to the server.
If you are managing an ODocumentDatabase you have to call the incrementalBackup()
method that accepts a String path parameter to the backup directory:
ODatabaseDocumentTx documentDatabase = new ODatabaseDocumentTx(dbURL);
documentDatabase.open("root", "password");
documentDatabase.incrementalBackup("/tmp/backup");
If you are using the OrientGraph interface you have to get the raw graph before calling the incrementalBackup()
method:
OrientGraph graphDatabase = new OrientGraphNoTx(dbURL);
graphDatabase.open("root", "password");
graphDatabase.getRawGraph().incrementalBackup("/tmp/backup");
Executing an Incremental Restore
Incremental Restore via the Console
Restore Database console command automatically recognizes if a backup contains incremental data. Restoring an incremental backup creates a new database with the restored content. You cannot perform an in-place incremental restore on an existing database. The execution of the create database command with the option -restore
builds a fresh database and performs the incremental restore starting from the backup path.
Example:
orientdb> create database remote:localhost/mydb root root plocal graph -restore=/tmp/backup
Creating database [remote:localhost/mydb] using the storage type [plocal]...
Connecting to database [remote:localhost/mydb] with user 'admin'...OK
Database created successfully.
Current database is: remote:localhost/mydb
In distributed mode restore will success only with a single node in the cluster. If you have 2 nodes or more in your cluster you have to use the standard restore procedure.
Incremental Restore via the Java API
You can perform an incremental restore via the Java API too.
To create a database from an incremental backup you can call from Java ODatabase#create(path-to-incremental-backup-directory)
.
Incremental Restore in Distributed Architecture
The incremental restore affects only the local node where the restore command is executed.
Let's suppose we have 3 nodes and we execute an incremental restore on node1. If we execute an incremental restore on node1 a new fresh database is created on all the 3 nodes, but only on node1 the restore procedure is performed. Thus we obtain the database correctly restored on node1 but an empty database on node2 and node 3.
You can overcome this inconsistency by executing a shutdown on all the nodes of the cluster not involved in the restore procedure (node2 and node3 in our example), so once restarted they will get the full database from node1.
Distributed Architecture
The incremental backup is used in the Distributed Architecture when a server node restarts. This avoids having to backup and transfer the entire database across the network.
Internals
File Format
In case of incremental backup, the content of the zip file is not the database directory, but rather meta files needed to update the database with the delta. Example of the content:
- Employee.pcl
- Person.pcl.incremental
- Person.pcm.incremental
This means only three files are changed. Employee.pcl is a full file, while the other two files with extension ".incremental" are incremental. Incremental files contain all the page changes and have the following format:
+----------------+-----------------+
| PAGE NUMBER | PAGE CONTENT |
| (long) | byte[] |
+----------------+-----------------+
Export and Import
OrientDB supports export and import operations, like any database management system.
This Section provides information on how to export-to and import-from JSON, as well on how to import from a RDBMS and Neo4j.
Neo4j is a registered trademark of Neo Technology, Inc.
Export to & Import from JSON
The EXPORT DATABASE
command exports the current opened database into a file. The exported file is in the Export JSON format. By default, it compresses the file using the GZIP algorithm.
Using exports with the IMPORT DATABASE
command, you can migrate the database between different releases of OrientDB without losing data. When doing this, if you receive an error relating to the database version, export the database using the same version of OrientDB on which you created the database.
orientdb> EXPORT DATABASE /temp/petshop.export
Exporting current database to: /temp/petshop.export...
Exporting database info...OK
Exporting dictionary...OK
Exporting schema...OK
Exporting clusters...
- Exporting cluster 'metadata' (records=11) -> ...........OK
- Exporting cluster 'index' (records=0) -> OK
- Exporting cluster 'default' (records=779) -> OK
- Exporting cluster 'csv' (records=1000) -> OK
- Exporting cluster 'binary' (records=1001) -> OK
- Exporting cluster 'person' (records=7) -> OK
- Exporting cluster 'animal' (records=5) -> OK
- Exporting cluster 'animalrace' (records=0) -> OK
- Exporting cluster 'animaltype' (records=1) -> OK
- Exporting cluster 'orderitem' (records=0) -> OK
- Exporting cluster 'order' (records=0) -> OK
- Exporting cluster 'city' (records=3) -> OK
Export of database completed.
Exports versus Backups
Exports don't lock the database. Instead, they browse the contents. This means that OrientDB can execute concurrent operations during the export, but the exported database may not be an exact replica from the time when you issued the command. If you need a database snapshot, use backups.
The BACKUP DATABASE
command does create a consistent copy of the database, but it locks the database. During the backup, the database remains in read-only mode, all concurrent write operations are blocked until the backup finishes. In the event that you need a database snapshot and the ability to perform read/write operations during the backup, set up a distributed cluster of nodes.
NOTE: Even though the export file is 100% JSON, there are some constraints in the JSON format, where the field order must be kept. Modifying the file to adjust the indentation may make the file unusable in database imports.
Importing Databases
Once you have exported your database, you can import it using the IMPORT DATABASE
command.
orientdb> IMPORT DATABASE /temp/petshop.export.gz -preserveClusterIDs=true
Importing records...
- Imported records into the cluster 'internal': 5 records
- Imported records into the cluster 'index': 4 records
- Imported records into the cluster 'default': 1022 records
- Imported records into the cluster 'orole': 3 records
- Imported records into the cluster 'ouser': 3 records
- Imported records into the cluster 'csv': 100 records
- Imported records into the cluster 'binary': 101 records
- Imported records into the cluster 'account': 1005 records
- Imported records into the cluster 'company': 9 records
- Imported records into the cluster 'profile': 9 records
- Imported records into the cluster 'whiz': 1000 records
- Imported records into the cluster 'address': 164 records
- Imported records into the cluster 'city': 55 records
- Imported records into the cluster 'country': 55 records
- Imported records into the cluster 'animalrace': 3 records
- Imported records into the cluster 'ographvertex': 102 records
- Imported records into the cluster 'ographedge': 101 records
- Imported records into the cluster 'graphcar': 1 records
For more information, see
Export Format
When you run the EXPORT DATABASE
command, OrientDB exports the database into a zipped file using a special JSON format. When you run the IMPORT DATABASE
command, OrientDB unzips the file and parses the JSON, making the import.
Sections
Export files for OrientDB use the following sections. Note that while the export format is 100% JSON, there are some constraints in the format, where the field order must be kept. Additionally, modifying the file to adjust the indentation (as has been done in the examples below), may make it unusable in database imports.
Info Section
The first section contains the resuming database information as well as all versions used during the export. OrientDB uses this information to check for compatibility during the import.
"info": {
"name": "demo",
"default-cluster-id": 2,
"exporter-format": 2,
"engine-version": "1.7-SNAPSHOT",
"storage-config-version": 2,
"schema-version": 4,
"mvrbtree-version": 0
}
Parameter | Description | JSON Type |
---|---|---|
"name" | Defines the name of the database. | String |
"default-cluster-id" | Defines the Cluster ID to use by default. Range: 0-32,762. | Integer |
"exporter-format" | Defines the version of the database exporter. | Integer |
"engine-version" | Defines the version of OrientDB. | String |
"storage-version" | Defines the version of the Storage layer. | Integer |
"schema-version" | Defines the version of the schema exporter. | Integer |
"mvrbtree-version" | Defines the version of the MVRB-Tree. | Integer |
Clusters Section
This section defines the database structure in clusters. It is formed from a list with an entry for each cluster in the database.
"clusters": [
{"name": "internal", "id": 0, "type": "PHYSICAL"},
{"name": "index", "id": 1, "type": "PHYSICAL"},
{"name": "default", "id": 2, "type": "PHYSICAL"}
]
Parameter | Description | JSON Type |
---|---|---|
"name" | Defines the logical name of the cluster. | String |
"id" | Defines the Cluster ID. Range: 0-32, 767. | Integer |
"type" | Defines the cluster type: PHYSICAL , LOGICAL and MEMORY . | String |
Schema Section
This section defines the database schema as classes and properties.
"schema":{
"version": 210,
"classes": [
{"name": "Account", "default-cluster-id": 9, "cluster-ids": [9],
"properties": [
{"name": "binary", "type": "BINARY", "mandatory": false, "not-null": false},
{"name": "birthDate", "type": "DATE", "mandatory": false, "not-null": false},
{"name": "id", "type": "INTEGER", "mandatory": false, "not-null": false}
]
}
]
}
Parameter | Description | JSON Type |
---|---|---|
"version" | Defines the version of the record storing the schema. Range: 0-2,147,483,647. | Integer |
"classes" | Defines a list of entries for each class in the schema. | Array |
Parameters for the Classes Subsection:
Parameter | Description | JSON Type |
---|---|---|
"name" | Defines the logical name of the class. | String |
"default-cluster-id" | Defines the default Cluster ID for the class. It represents the cluster that stores the class records. | Integer |
"cluster-ids" | Defines an array of Cluster ID's that store the class records. The first ID is always the default Cluster ID. | Array of Integers |
"properties" | Defines a list of entries for each property for the class in the schema. | Array |
Parameters for the Properties Sub-subsection:
Parameter | Description | JSON Type |
---|---|---|
"name" | Defines the logical name of the property. | String |
"type" | Defines the property type. | String |
"mandatory" | Defines whether the property is mandatory. | Boolean |
"not-null" | Defines whether the property accepts a NULL value. | Boolean |
Records Section
This section defines the exported record with metadata and fields. Entries for metadata are distinguished from fields by the @
symbol.
"records": [
{"@type": "d", "@rid": "#12:476", "@version": 0, "@class": "Account",
"account_id": 476,
"date": "2011-12-09 00:00:00:0000",
"@fieldTypes": ["account_id=i", "date=t"]
},
{"@type": "d", "@rid": "#12:477", "@version": 0, "@class": "Whiz",
"id": 477,
"date": "2011-12-09 00:00:00:000",
"text": "He in office return He inside electronics for $500,000 Jay",
"@fieldTypes": "date=t"
}
]
Parameters for Metadata
Parameter | Description | JSON Type |
---|---|---|
"@type" | Defines the record-type: d for Document, b for Binary. | String |
"@rid" | Defines the Record ID, using the format: <cluster-id>:<cluster-position> . | String |
"@version" | Defines the record version. Range: 0-2, 147, 483, 647. | Integer |
"@class" | Defines the logical class name for the record. | String |
"@fieldTypes" | Defines an array of the types for each field in this record. | Any |
Supported Field Types
Value | Type |
---|---|
l | Long |
f | Float |
d | Double |
s | Short |
t | Datetime |
d | Date |
c | Decimal |
b | Byte |
Full Example
{
"info":{
"name": "demo",
"default-cluster-id": 2,
"exporter-version": 2,
"engine-version": "1.0rc8-SNAPSHOT",
"storage-config-version": 2,
"schema-version": 4,
"mvrbtree-version": 0
},
"clusters": [
{"name": "internal", "id": 0, "type": "PHYSICAL"},
{"name": "index", "id": 1, "type": "PHYSICAL"},
{"name": "default", "id": 2, "type": "PHYSICAL"},
{"name": "orole", "id": 3, "type": "PHYSICAL"},
{"name": "ouser", "id": 4, "type": "PHYSICAL"},
{"name": "orids", "id": 5, "type": "PHYSICAL"},
{"name": "csv", "id": 6, "type": "PHYSICAL"},
{"name": "binary", "id": 8, "type": "PHYSICAL"},
{"name": "account", "id": 9, "type": "PHYSICAL"},
{"name": "company", "id": 10, "type": "PHYSICAL"},
{"name": "profile", "id": 11, "type": "PHYSICAL"},
{"name": "whiz", "id": 12, "type": "PHYSICAL"},
{"name": "address", "id": 13, "type": "PHYSICAL"},
{"name": "city", "id": 14, "type": "PHYSICAL"},
{"name": "country", "id": 15, "type": "PHYSICAL"},
{"name": "dummy", "id": 16, "type": "PHYSICAL"},
{"name": "ographvertex", "id": 26, "type": "PHYSICAL"},
{"name": "ographedge", "id": 27, "type": "PHYSICAL"},
{"name": "graphvehicle", "id": 28, "type": "PHYSICAL"},
{"name": "graphcar", "id": 29, "type": "PHYSICAL"},
{"name": "graphmotocycle", "id": 30, "type": "PHYSICAL"},
{"name": "newv", "id": 31, "type": "PHYSICAL"},
{"name": "mappoint", "id": 33, "type": "PHYSICAL"},
{"name": "person", "id": 35, "type": "PHYSICAL"},
{"name": "order", "id": 36, "type": "PHYSICAL"},
{"name": "post", "id": 37, "type": "PHYSICAL"},
{"name": "comment", "id": 38, "type": "PHYSICAL"}
],
"schema":{
"version": 210,
"classes": [
{"name": "Account", "default-cluster-id": 9, "cluster-ids": [9],
"properties": [
{"name": "binary", "type": "BINARY", "mandatory": false, "not-null": false},
{"name": "birthDate", "type": "DATE", "mandatory": false, "not-null": false},
{"name": "id", "type": "INTEGER", "mandatory": false, "not-null": false}
]
},
{"name": "Address", "default-cluster-id": 13, "cluster-ids": [13]
},
{"name": "Animal", "default-cluster-id": 17, "cluster-ids": [17]
},
{"name": "AnimalRace", "default-cluster-id": 18, "cluster-ids": [18]
},
{"name": "COMMENT", "default-cluster-id": 38, "cluster-ids": [38]
},
{"name": "City", "default-cluster-id": 14, "cluster-ids": [14]
},
{"name": "Company", "default-cluster-id": 10, "cluster-ids": [10], "super-class": "Account",
"properties": [
]
},
{"name": "Country", "default-cluster-id": 15, "cluster-ids": [15]
},
{"name": "Dummy", "default-cluster-id": 16, "cluster-ids": [16]
},
{"name": "GraphCar", "default-cluster-id": 29, "cluster-ids": [29], "super-class": "GraphVehicle",
"properties": [
]
},
{"name": "GraphMotocycle", "default-cluster-id": 30, "cluster-ids": [30], "super-class": "GraphVehicle",
"properties": [
]
},
{"name": "GraphVehicle", "default-cluster-id": 28, "cluster-ids": [28], "super-class": "OGraphVertex",
"properties": [
]
},
{"name": "MapPoint", "default-cluster-id": 33, "cluster-ids": [33],
"properties": [
{"name": "x", "type": "DOUBLE", "mandatory": false, "not-null": false},
{"name": "y", "type": "DOUBLE", "mandatory": false, "not-null": false}
]
},
{"name": "OGraphEdge", "default-cluster-id": 27, "cluster-ids": [27], "short-name": "E",
"properties": [
{"name": "in", "type": "LINK", "mandatory": false, "not-null": false, "linked-class": "OGraphVertex"},
{"name": "out", "type": "LINK", "mandatory": false, "not-null": false, "linked-class": "OGraphVertex"}
]
},
{"name": "OGraphVertex", "default-cluster-id": 26, "cluster-ids": [26], "short-name": "V",
"properties": [
{"name": "in", "type": "LINKSET", "mandatory": false, "not-null": false, "linked-class": "OGraphEdge"},
{"name": "out", "type": "LINKSET", "mandatory": false, "not-null": false, "linked-class": "OGraphEdge"}
]
},
{"name": "ORIDs", "default-cluster-id": 5, "cluster-ids": [5]
},
{"name": "ORole", "default-cluster-id": 3, "cluster-ids": [3],
"properties": [
{"name": "mode", "type": "BYTE", "mandatory": false, "not-null": false},
{"name": "name", "type": "STRING", "mandatory": true, "not-null": true},
{"name": "rules", "type": "EMBEDDEDMAP", "mandatory": false, "not-null": false, "linked-type": "BYTE"}
]
},
{"name": "OUser", "default-cluster-id": 4, "cluster-ids": [4],
"properties": [
{"name": "name", "type": "STRING", "mandatory": true, "not-null": true},
{"name": "password", "type": "STRING", "mandatory": true, "not-null": true},
{"name": "roles", "type": "LINKSET", "mandatory": false, "not-null": false, "linked-class": "ORole"}
]
},
{"name": "Order", "default-cluster-id": 36, "cluster-ids": [36]
},
{"name": "POST", "default-cluster-id": 37, "cluster-ids": [37],
"properties": [
{"name": "comments", "type": "LINKSET", "mandatory": false, "not-null": false, "linked-class": "COMMENT"}
]
},
{"name": "Person", "default-cluster-id": 35, "cluster-ids": [35]
},
{"name": "Person2", "default-cluster-id": 22, "cluster-ids": [22],
"properties": [
{"name": "age", "type": "INTEGER", "mandatory": false, "not-null": false},
{"name": "firstName", "type": "STRING", "mandatory": false, "not-null": false},
{"name": "lastName", "type": "STRING", "mandatory": false, "not-null": false}
]
},
{"name": "Profile", "default-cluster-id": 11, "cluster-ids": [11],
"properties": [
{"name": "hash", "type": "LONG", "mandatory": false, "not-null": false},
{"name": "lastAccessOn", "type": "DATETIME", "mandatory": false, "not-null": false, "min": "2010-01-01 00:00:00"},
{"name": "name", "type": "STRING", "mandatory": false, "not-null": false, "min": "3", "max": "30"},
{"name": "nick", "type": "STRING", "mandatory": false, "not-null": false, "min": "3", "max": "30"},
{"name": "photo", "type": "TRANSIENT", "mandatory": false, "not-null": false},
{"name": "registeredOn", "type": "DATETIME", "mandatory": false, "not-null": false, "min": "2010-01-01 00:00:00"},
{"name": "surname", "type": "STRING", "mandatory": false, "not-null": false, "min": "3", "max": "30"}
]
},
{"name": "PropertyIndexTestClass", "default-cluster-id": 21, "cluster-ids": [21],
"properties": [
{"name": "prop1", "type": "STRING", "mandatory": false, "not-null": false},
{"name": "prop2", "type": "INTEGER", "mandatory": false, "not-null": false},
{"name": "prop3", "type": "BOOLEAN", "mandatory": false, "not-null": false},
{"name": "prop4", "type": "INTEGER", "mandatory": false, "not-null": false},
{"name": "prop5", "type": "STRING", "mandatory": false, "not-null": false}
]
},
{"name": "SQLDropIndexTestClass", "default-cluster-id": 23, "cluster-ids": [23],
"properties": [
{"name": "prop1", "type": "DOUBLE", "mandatory": false, "not-null": false},
{"name": "prop2", "type": "INTEGER", "mandatory": false, "not-null": false}
]
},
{"name": "SQLSelectCompositeIndexDirectSearchTestClass", "default-cluster-id": 24, "cluster-ids": [24],
"properties": [
{"name": "prop1", "type": "INTEGER", "mandatory": false, "not-null": false},
{"name": "prop2", "type": "INTEGER", "mandatory": false, "not-null": false}
]
},
{"name": "TestClass", "default-cluster-id": 19, "cluster-ids": [19],
"properties": [
{"name": "name", "type": "STRING", "mandatory": false, "not-null": false},
{"name": "testLink", "type": "LINK", "mandatory": false, "not-null": false, "linked-class": "TestLinkClass"}
]
},
{"name": "TestLinkClass", "default-cluster-id": 20, "cluster-ids": [20],
"properties": [
{"name": "testBoolean", "type": "BOOLEAN", "mandatory": false, "not-null": false},
{"name": "testString", "type": "STRING", "mandatory": false, "not-null": false}
]
},
{"name": "Whiz", "default-cluster-id": 12, "cluster-ids": [12],
"properties": [
{"name": "account", "type": "LINK", "mandatory": false, "not-null": false, "linked-class": "Account"},
{"name": "date", "type": "DATE", "mandatory": false, "not-null": false, "min": "2010-01-01"},
{"name": "id", "type": "INTEGER", "mandatory": false, "not-null": false},
{"name": "replyTo", "type": "LINK", "mandatory": false, "not-null": false, "linked-class": "Account"},
{"name": "text", "type": "STRING", "mandatory": true, "not-null": false, "min": "1", "max": "140"}
]
},
{"name": "classclassIndexManagerTestClassTwo", "default-cluster-id": 25, "cluster-ids": [25]
},
{"name": "newV", "default-cluster-id": 31, "cluster-ids": [31], "super-class": "OGraphVertex",
"properties": [
{"name": "f_int", "type": "INTEGER", "mandatory": false, "not-null": false}
]
},
{"name": "vertexA", "default-cluster-id": 32, "cluster-ids": [32], "super-class": "OGraphVertex",
"properties": [
{"name": "name", "type": "STRING", "mandatory": false, "not-null": false}
]
},
{"name": "vertexB", "default-cluster-id": 34, "cluster-ids": [34], "super-class": "OGraphVertex",
"properties": [
{"name": "map", "type": "EMBEDDEDMAP", "mandatory": false, "not-null": false},
{"name": "name", "type": "STRING", "mandatory": false, "not-null": false}
]
}
]
},
"records": [{
"@type": "d", "@rid": "#12:476", "@version": 0, "@class": "Whiz",
"id": 476,
"date": "2011-12-09 00:00:00:000",
"text": "Los a went chip, of was returning cover, In the",
"@fieldTypes": "date=t"
},{
"@type": "d", "@rid": "#12:477", "@version": 0, "@class": "Whiz",
"id": 477,
"date": "2011-12-09 00:00:00:000",
"text": "He in office return He inside electronics for $500,000 Jay",
"@fieldTypes": "date=t"
}
]
}
Import from RDBMS
NOTE: As of OrientDB 2.0, you can use the OrientDB-ETL module to import data from an RDBMS. You can use ETL also with 1.7.x by installing it as a separate module.
OrientDB supports a subset of SQL, so importing a database created as "Relational" is straightforward. For the sake of simplicity, consider your Relational database having just these two tables:
- POST
- COMMENT
Where the relationship is between Post and comment as One-2-Many.
TABLE POST:
+----+----------------+
| id | title |
+----+----------------+
| 10 | NoSQL movement |
| 20 | New OrientDB |
+----+----------------+
TABLE COMMENT:
+----+--------+--------------+
| id | postId | text |
+----+--------+--------------+
| 0 | 10 | First |
| 1 | 10 | Second |
| 21 | 10 | Another |
| 41 | 20 | First again |
| 82 | 20 | Second Again |
+----+--------+--------------+
- Import using the Document Model (relationships as links)
- Import using the Graph Model (relationships as edges)
Import from a Relational Database
Relational databases typically query and manipulate data with SQL. Given that OrientDB supports a subset of SQL, it is relatively straightforward to import data from a Relational databases to OrientDB. You can manage imports using the Java API, OrientDB Studio or the OrientDB Console. The examples below use the Console.
This guide covers importing into the Document Model. Beginning with version 2.0, you can import into the Graph Model using the ETL Module. From version 1.7.x you can still use ETL by installing it as a separate module
For these examples, assume that your Relational database, (referred to as reldb
in the code), contains two tables: Post
and Comment
. The relationship between these tables is one-to-many.
reldb>SELECT * FROM post;
+----+----------------+ | id | title | +----+----------------+ | 10 | NoSQL movement | | 20 | New OrientDB | +----+----------------+ reldb>SELECT * FROM comment;
+----+--------+--------------+ | id | postId | text | +----+--------+--------------+ | 0 | 10 | First | | 1 | 10 | Second | | 21 | 10 | Another | | 41 | 20 | First again | | 82 | 20 | Second Again | +----+--------+--------------+
Given that the Relational Model doesn't use concepts from Object Oriented Programming, there are some things to consider in the transition from a Relational database to OrientDB.
-
In Relational databases there is no concept of class, so in the import to OrientDB you need to create on class per table.
-
In Relational databases, one-to-many references invert from the target table to the source table.
Table POST <- (foreign key) Table COMMENT
In OrientDB, it follows the Object Oriented Model, so you have a collection of links connecting instances of
Post
andComment
.Class POST ->* (collection of links) Class COMMENT
Exporting Relational Databases
Most Relational database management systems provide a way to export the database into SQL format. What you specifically need from this is a text file that contains the SQL INSERT
commands to recreate the database from scratch. For example,
- MySQL: the
mysqldump
utility. - Oracle Database: the Datapump utilities.
- Microsoft SQL Server: the Import and Export Wizard.
When you run this utility on the example database, it produces an .sql
file that contains the exported SQL of the Relational database.
DROP TABLE IF EXISTS post;
CREATE TABLE post (
id INT(11) NOT NULL AUTO_INCREMENT,
title VARCHAR(128),
PRIMARY KEY (id)
);
DROP TABLE IF EXISTS comment;
CREATE TABLE comment (
id INT(11) NOT NULL AUTO_INCREMENT,
postId INT(11),
text TEXT,
PRIMARY KEY (id),
CONSTRAINT `fk_comments`
FOREIGN KEY (`postId` )
REFERENCES `post` (`id` )
);
INSERT INTO POST (id, title) VALUES( 10, 'NoSQL movement' );
INSERT INTO POST (id, title) VALUES( 20, 'New OrientDB' );
INSERT INTO COMMENT (id, postId, text) VALUES( 0, 10, 'First' );
INSERT INTO COMMENT (id, postId, text) VALUES( 1, 10, 'Second' );
INSERT INTO COMMENT (id, postId, text) VALUES( 21, 10, 'Another' );
INSERT INTO COMMENT (id, postId, text) VALUES( 41, 20, 'First again' );
INSERT INTO COMMENT (id, postId, text) VALUES( 82, 20, 'Second Again' );
Modifying the Export File
Importing from the Relational database requires that you modify the SQL file to make it usable by OrientDB. In order to do this, you need to open the SQL file, (called export.sql
below), in a text editor and modify the commands there. Once this is done, you can execute the file on the Console using batch mode.
Database
In order to import a data into OrientDB, you need to have a database ready to receive the import. Note that the example export.sql
file doesn't include statements to create the database. You can either create a new database or use an existing one.
Using New Databases
In creating a database for the import, you can either create a volatile in-memory database, (one that is only available while OrientDB is running), or you can create a persistent disk-based database. For a persistent database, you can create it on a remote server or locally through the PLocal mode.
The recommended method is PLocal, given that it offers better performance on massive inserts.
-
Using the embedded Plocal mode:
$
vim export.sql
CREATE DATABASE PLOCAL:/tmp/db/blog admin_user admin_passwd PLOCAL DOCUMENT
Here, the
CREATE DATABASE
command creates a new database at/tmp/db/blog
. -
Using the Remote mode:
$
vim export.sql
CREATE DATABASE REMOTE:localhost/blog root_user dkdf383dhdsj PLOCAL DOCUMENT
This creates a database at the URL
http://localhost/blog
.
NOTE: When you create remote databases, you need the server credentials to access it. The user
root
and its password are stored in the$ORIENTDB_HOME/config/orientdb-server-config.xml
configuration file.
Using Existing Databases
In the event that you already have a database set up and ready for the import, instead of creating a database add a line that connects to that databases, using the CONNECT
command.
-
Using the embedded PLocal mode:
$
vim export.sh
CONNECT PLOCAL:/tmp/db/blog admin_user admin_passwd
This connects to the database at
/tmp/db/blog
. -
Using the Remote mode:
$
vim export.sql
CONNECT REMOTE:localhost/blog admin_user admin_passwd
This connects to the database at the URL
http://localhost/blog
.
Declaring Intent
In the SQL file, after you create or connect to the database, you need to declare your intention to perform a massive insert. Intents allow you to utilize automatic tuning OrientDB for maximum performance on particular operations, such as large inserts or reads.
$vim export.sh
...DECLARE INTENT MASSIVEINSERT
Creating Classes
Relational databases have no parallel to concepts in Object Oriented programming, such as classes. Conversely, OrientDB doesn't have a concept of tables in the Relational sense.
Modify the SQL file, changing CREATE TABLE
statements to CREATE CLASS
commands:
$vim export.sql
...CREATE CLASS Post CREATE CLASS Comment
NOTE: In cases where your Relational database was created using Object Relational Mapping, or ORM, tools, such as Hibernate or Data Nucleus, you have to rebuild the original Object Oriented Structure directly in OrientDB.
Create Links
In the Relational database, the relationship between the post
and comment
was handled through foreign keys on the id
fields. OrientDB handles relationships differently, using links between two or more records of the Document type.
By default, the CREATE LINK
command creates a direct relationship in your object model. Navigation goes from Post
to Comment
and not vice versa, which is the case for the Relational database. You'll need to use the INVERSE
keyword to make the links work in both directions.
Add the following line after the INSERT
statements.
$vim export.sql
...CREATE LINK comments TYPE LINKSET FROM comment.postId TO post.id INVERSE
Remove Constraints
Unlike how Relational databases handle tables, OrientDB does not require you to create a strict schema on your classes. The properties on each class are defined through the INSERT
statements. That is, id
and title
on Post
and id
, postId
and text
on Comment
.
Given that you created a link in the above section, the property postId
is no longer necessary. Instead of modifying each INSERT
statement, you can use the UPDATE
command to remove them at the end:
$vim export.sql
...UPDATE comment REMOVE postId
Bear in mind, this is an optional step. The database will still function if you leave this field in place.
Expected Output
When you've finished, remove any statements that OrientDB does not support. With the changes above this leaves you with a file similar to the one below:
$ cat export.sql
CONNECT plocal:/tmp/db/blog admin admin
DECLARE INTENT MASSIVEINSERT
CREATE CLASS Post
CREATE CLASS Comment
INSERT INTO Post (id, title) VALUES( 10, 'NoSQL movement' )
INSERT INTO Post (id, title) VALUES( 20, 'New OrientDB' )
INSERT INTO Comment (id, postId, text) VALUES( 0, 10, 'First' )
INSERT INTO Comment (id, postId, text) VALUES( 1, 10, 'Second' )
INSERT INTO Comment (id, postId, text) VALUES( 21, 10, 'Another' )
INSERT INTO Comment (id, postId, text) VALUES( 41, 20, 'First again' )
INSERT INTO Comment (id, postId, text) VALUES( 82, 20, 'Second Again' )
CREATE LINK comments TYPE LINKSET FROM Comment.postId TO Post.id INVERSE
UPDATE Comment REMOVE postId
Importing Databases
When you finish modifying the SQL file, you can execute it through the Console in batch mode. This is done by starting the Console with the SQL file given as the first argument.
$ $ORIENTDB_HOME/bin/console.sh export.sql
When the OrientDB starts, it executes each of the commands given in the SQL files, creating or connecting to the database, creating the classes and inserting the data from the Relational database. You now have a working instance of OrientDB to use.
Using the Database
You now have an OrientDB Document database where relationships are direct and handled without the use of joins.
-
Query for all posts with comments:
orientdb>
SELECT FROM Post WHERE comments.size() > 0
-
Query for all posts where the comments contain the word "flame" in the
text
property:orientdb>
SELECT FROM Post WHERE comments CONTAINS(text LIKE '%flame%')
-
Query for all posts with comments made today, assuming that you have added a
date
property to theComment
class:orientdb>
SELECT FROM Post WHERE comments CONTAINS(date > '2011-04-14 00:00:00')
For more information, see
Import from RDBMS to Graph Model
To import from RDBMS to OrientDB using the Graph Model the ETL tool is the suggested way to do it. Take a look at: Import from CSV to a Graph.
Import from Neo4j
Neo4j is an open-source graph database that queries and manipulates data using its own Cypher Query Language.
For more information on the differences between Neo4j and OrientDB, please refer to the OrientDB vs. Neo4j page.
Neo4j and Cypher are registered trademark of Neo Technology, Inc.
Migration Strategies
Importing data from Neo4j into OrientDB is a straightforward process.
To migrate, you may use the Neo4j to OrientDB Importer. Starting from OrientDB version 2.2, this is the preferred way to migrate from Neo4j.
The Neo4j to OrientDB Importer allows you to migrate Neo4j's nodes, relationships, constraints and indexes. For more details, please refer to the Neo4j to OrientDB Importer Section.
Note: if your data is in CSV format, you can migrate to OrientDB using the OrientDB's ETL tool.
This is a legacy strategy to migrate from Neo4j. The new strategy is to migrate using the Neo4j to OrientDB Importer. |
Import from Neo4j using GraphML
This section describes the process of importing data from Neo4j to OrientDB using GraphML. For general information on the possible Neo4j to OrientDB migration strategies, please refer to the Import from Neo4j section.
Neo4j can export in GraphML, an XML-based file format for graphs. Given that OrientDB can read GraphML, you can use this file format to import data from Neo4j into OrientDB, using the Console or the Java API.
Note:
For large and complex datasets, the preferred way to migrate from Neo4j is using the Neo4j to OrientDB Importer.
Neo4j and Cypher are registered trademark of Neo Technology, Inc.
Exporting GraphML
In order to export data from Neo4j into GraphML, you need to install the Neo4j Shell Tools plugin. Once you have this package installed, you can use the export-graphml
utility to export the database.
-
Change into the Neo4j home directory:
$
cd /path/to/neo4j-community-2.3.2
-
Download the Neo4j Shell Tools:
$
curl http://dist.neo4j.org/jexp/shell/neo4j-shell-tools_2.3.2.zip \ -o neo4j-shell-tools.zip
-
Unzip the
neo4j-shell-tools.zip
file into thelib
directory:$
unzip neo4j-shell-tools.zip -d lib
-
Restart the Neo4j Server. In the event that it's not running,
start
it:$
./bin/neo4j restart
-
Once you have Neo4j restarted with the Neo4j Shell Tools, launch the Neo4j Shell tool, located in the
bin/
directory:$
./bin/neo4j-shell
Welcome to the Neo4j Shell! Enter 'help' for a list of commands NOTE: Remote Neo4j graph database service 'shell' at port 1337 neo4j-sh (0)$ -
Export the database into GraphML:
neo4j-sh (0)$
export-graphml -t -o /tmp/out.graphml
Wrote to GraphML-file /tmp/out.graphml 0. 100%: nodes = 302 rels = 834 properties = 4221 time 59 sec total 59 sec
This exports the database to the path /tmp/out.graphml
.
Importing GraphML
There are three methods available in importing the GraphML file into OrientDB: through the Console, through Gremlin or through the Java API.
Importing through the OrientDB Console
For more recent versions of OrientDB, you can import data from GraphML through the OrientDB Console. If you have version 2.0 or greater, this is the recommended method given that it can automatically translate the Neo4j labels into classes.
-
Log into the OrientDB Console.
$
$ORIENTDB_HOME/bin/console.sh
-
In OrientDB, create a database to receive the import:
orientdb>
CREATE DATABASE PLOCAL:/tmp/db/test
Creating database [plocal:/tmp/db/test] using the storage type [plocal]... Database created successfully. Current database is: plocal:/tmp/db/test -
Import the data from the GraphML file:
orientdb {db=test}>
IMPORT DATABASE /tmp/out.graphml
Importing GRAPHML database database from /tmp/out.graphml... Transaction 8 has been committed in 12ms
This imports the Neo4j database into OrientDB on the test
database.
Importing through the Gremlin Console
For older versions of OrientDB, you can import data from GraphML through the Gremlin Console. If you have a version 1.7 or earlier, this is the method to use. It is not recommended on more recent versions, given that it doesn't consider labels declared in Neo4j. In this case, everything imports as the base vertex and edge classes, (that is, V
and E
). This means that, after importing through Gremlin you need to refactor you graph elements to fit a more structured schema.
To import the GraphML file into OrientDB, complete the following steps:
-
Launch the Gremlin Console:
$
$ORIENTDB_HOME/bin/gremlin.sh
\,,,/ (o o) -----oOOo-(_)-oOOo----- -
From the Gremlin Console, create a new graph, specifying the path to your Graph database, (here
/tmp/db/test
):gremlin>
g = new OrientGraph("plocal:/tmp/db/test");
==>orientgraph[plocal:/db/test] -
Load the GraphML file into the graph object (that is,
g
):gremlin>
g.loadGraphML("/tmp/out.graphml");
==>null -
Exit the Gremlin Console:
gremlin>
quit
This imports the GraphML file into your OrientDB database.
Importing through the Java API
OrientDB Console calls the Java API. Using the Java API directly allows you greater control over the import process. For instance,
new OGraphMLReader(new OrientGraph("plocal:/temp/bettergraph")).inputGraph("/temp/neo4j.graphml");
This line imports the GraphML file into OrientDB.
Defining Custom Strategies
Beginning in version 2.1, OrientDB allows you to modify the import process through custom strategies for vertex and edge attributes. It supports the following strategies:
com.orientechnologies.orient.graph.graphml.OIgnoreGraphMLImportStrategy
Defines attributes to ignore.com.orientechnologies.orient.graph.graphml.ORenameGraphMLImportStrategy
Defines attributes to rename.
Examples
-
Ignore the vertex attribute
type
:new OGraphMLReader(new OrientGraph("plocal:/temp/bettergraph")).defineVertexAttributeStrategy("__type__", new OIgnoreGraphMLImportStrategy()).inputGraph("/temp/neo4j.graphml");
-
Ignore the edge attribute
weight
:new OGraphMLReader(new OrientGraph("plocal:/temp/bettergraph")).defineEdgeAttributeStrategy("weight", new OIgnoreGraphMLImportStrategy()).inputGraph("/temp/neo4j.graphml");
-
Rename the vertex attribute
type
in justtype
:new OGraphMLReader(new OrientGraph("plocal:/temp/bettergraph")).defineVertexAttributeStrategy("__type__", new ORenameGraphMLImportStrategy("type")).inputGraph("/temp/neo4j.graphml");
Import Tips and Tricks
Dealing with Memory Issues
In the event that you experience memory issues while attempting to import from Neo4j, you might consider reducing the batch size. By default, the batch size is set to 1000
. Smaller value causes OrientDB to process the import in smaller units.
-
Import with adjusted batch size through the Console:
orientdb {db=test}>
IMPORT DATABASE /tmp/out.graphml batchSize=100
-
Import with adjusted batch size through the Java API:
new OGraphMLReader(new OrientGraph("plocal:/temp/bettergraph")).setBatchSize(100).inputGraph("/temp/neo4j.graphml");
Storing the Vertex ID's
By default, OrientDB updates the import to use its own ID's for vertices. If you want to preserve the original vertex ID's from Neo4j, use the storeVertexIds
option.
-
Import with the original vertex ID's through the Console:
orientdb {db=test}>
IMPORT DATABASE /tmp/out.graphml storeVertexIds=true
-
Import with the original vertex ID's through the Java API:
new OGraphMLReader(new OrientGraph("plocal:/temp/bettergraph")).setStoreVertexIds(true).inputGraph("/temp/neo4j.graphml");
Example
A complete example of a migration from Neo4j to OrientDB using the GraphML method can be found in the section Tutorial: Importing the movie Database from Neo4j.
Monitoring
JMX
Read Cache
JMX bean name: com.orientechnologies.orient.core.storage.cache.local:type=O2QCacheMXBean
It has following members:
usedMemory
,usedMemoryInMB
,usedMemoryInGB
which is amount of direct memory consumed by read cache in different units of measurementscacheHits
is percent of cases when records will be downloaded not from disk but from read cacheclearCacheStatistics()
method may be called to clear cache hits statics so we always may start to gather cache hits statistic from any moment of timeamSize
,a1OutSize
,a1InSize
is the size of LRU queues are used in 2Q algorithm
Write Cache
JMX bean name: com.orientechnologies.orient.core.storage.cache.local:type=OWOWCacheMXBean,name=<storage name>,id=<storage id>
Write cache alike read cache is not JVM wide, it is storage wide, but one JVM may run several servers and each server may contain storage with the same name, that is why we need such complex name.
JMX bean of write cache has following members:
writeCacheSize
,writeCacheSizeInMB
,writeCacheSizeInGB
provides size of data in different units which should be flushed to disk in background threadexclusiveWriteCacheSize
,exclusiveWriteCacheSizeInMB
,exclusiveWriteCacheSizeInGB
provides size of data which should be flushed to disk but contained only in write cache
More about memory model and data flow
At first when we read page we load it from disk and put it in read cache. Then we change page and put it back to read cache and write cache, but we do not copy page from read to write cache we merely send pointer to the same memory to write cache. Write cache flushes "dirty write page" in background thread. That is what property "writeCacheSize" shows us amount of data in dirty pages which should be flushed. But there are very rare situations when page which is rarely used still is not flushed on disk and read cache has not enough memory to keep it. In such case this page is removed from read cache , but pointer to this page still exists in write cache, that is what property "exclusiveWriteCacheSize" shows us. Please note that this value is more than 0 only during extremely high load.
The rest properties of write cache JMX bean are following:
lastFuzzyCheckpointDate
lastAmountOfFlushedPages
durationOfLastFlush
#Tools
Introduction to OrientDB Studio
People have different preferences to how they would like to interact with a database. Some prefer to work through an application, some through an API, some a console. OrientDB Studio is for those who are most comfortable operating on databases through a graphical user interface.
When you start the OrientDB Server, the JVM also runs Studio on port 2480. Running OrientDB on your local system, you can access it through the web browser by navigating to http://localhost:2480. To access Studio on a remote machine, you may want to use an SSH tunnel for better security.
From the Studio Home Page, you can:
- Connect to an existing database
- Create a new database
- Import a public (OrientDB) database
- Drop an existing database
- Start a migration to OrientDB using the tool Teleporter
- Access the Server Management features
Overview of Menus and Panels
This Section contains a list of all Studio's menu and a quick map between the menu and the corresponding panel in the program.
Horizontal Menus
Browse
If you click on the Browse menu the Browse panel will be shown.
The Browse panel includes an SQL query editor from which you can execute (and run the explain of) a query. Query results can be visualized in a table or JSON format. In case the executed query returns an editable recordset, you will be able to edit your data.
From this panel, returned recordsets can be sent to the Graph Editor to see their graph visualization.
See Also:
Schema
If you click on the Schema menu the Schema Manager will be shown.
The Schema Manager includes a lists of all Classes and Indexes of in the current database, and allows you create, edit or drop Classes, their Properties, and Indexes.
See Also:
Security
If you click on the Security menu the Security Manager will be shown.
The Security Manager includes a lists of all Users and Roles of in the current database, and allows you create, edit or drop Users and Roles.
See Also:
Graph
If you click on the Graph menu the Graph Editor will be shown.
The Graph Editor includes an SQL query editor from which you can execute a query to visualize a graph. It also allows you to modify a graph, by adding new vertices, edges and setting their properties.
See Also:
Functions
If you click on the Function menu the Function Management panel will be shown.
The Function Management panel allows you to create, edit and drop functions.
See Also:
DB
If you click on the DB menu the Database Management panel will be shown.
The Database Management panel includes several (structure and metadata) information about the current database. It also allows you to export the current database in JSON format.
See Also:
Vertical Menus
Dashboard
Servers Management
Cluster Management
Backup Management
Query Profiler
Security
Teleporter
Alerts Management
Common Database Operations
Connecting to an Existing Database
To connect to an existing database, select a database from the databases list and use a valid database user.
By default reader/reader can read records from the database, writer/writer can read, create, update and delete records. admin/admin has all rights.
Creating a New Database
To create a new database, click the "New DB" button from the Home Page:
Some information is needed to create a new database:
- Database name
- Database type (Document/Graph)
- Storage type (plocal/memory)
- Server user
- Server password
You can find the server credentials in the $ORIENTDB_HOME/config/orientdb-server-config.xml file:
<users>
<user name="root" password="pwd" resources="*" />
</users>
Once created, Studio will automatically login to the new database.
Importing a Public Database
Starting from version 2.2, Studio allows you to import databases from a public repository. These databases contain public data and bookmarked queries that will allow you to start playing with OrientDB and OrientDB SQL.
To install a public database, you will need the Server Credentials. Then, click the download button of the database that you are interested in. Then Studio will download and install in to your $ORIENTDB_HOME/databases directory. Once finished, Studio will automatically login to the newly installed database.
Dropping an Existing Database
To drop an existing database, select it from the databases list and click the trash icon. Studio will display a confirmation popup where you have to insert:
- Server User
- Server Password
and then click the "Drop database" button. You can find the server credentials in the $ORIENTDB_HOME/config/orientdb-server-config.xml file:
<users>
<user name="root" password="pwd" resources="*" />
</users>
search: keywords: ['Studio', 'database', 'Database Management']
Database Management Panel
This is the panel containing all the information about the current database.
#Structure Represents the database structure as clusters. Each cluster has the following information:
ID
, is the cluster IDName
, is the name of the clusterRecords
, are the total number of records stored in the clusterConflict Strategy
, is the conflict strategy used. I empty, the database's strategy is used as default
##Configuration
Contains the database configuration and custom properties. Here you can display and change the following settings:
dateFormat
, is the date format used in the database by default. Example: yyyy-MM-dddateTimeFormat
is the datetime format used in the database by default. Example: yyyy-MM-dd HH:mm:sslocaleCountry
, is the country used. "NO" means no country setlocaleLanguage
, is the language used. "no" means no language setcharSet
, is the charset used. Default is UTF-8timezone
, is the timezone used. Timezone is taken on database creationdefinitionVersion
, is the internal version used to store the metadataclusterSelection
, is the strategy used on selecting the cluster on creation of new record of a classminimumClusters
, minimum number of clusters to create when at class creationconflictStrategy
, is the database strategy for resolving conflicts
##Export Allows to export the current database in GZipped JSON format. To import the file into another database, use the Import Console Command.
Working with Data
Studio provides two main features to interact and work with data:
- The Browse panel
- The Graph Editor
Browse Panel
Studio supports auto recognition of the language you're using between those supported: SQL and Gremlin. While writing, use the auto-complete feature by pressing Ctrl + Space.
Other shortcuts are available in the query editor:
- Ctrl + Return to execute the query or just click the Run button
- Ctrl/Cmd + Z to undo changes
- Ctrl/Cmd + Shift + Z to redo changes
- Ctrl/Cmd + F to search in the editor
- Ctrl/Cmd + / to toggle a comment
Note: If you have multiple queries in the editor, you can select a single query with text selection and execute it with Ctrl + Return or the Run button
By clicking any @rid value in the result set, you will go into document edit mode if the record is a Document, otherwise you will go into vertex edit.
You can bookmark your queries by clicking the star icon in the results set or in the editor. To browse bookmarked queries, click the Bookmarks button. Studio will open the bookmarks list on the left, where you can edit/delete or rerun queries.
Studio saves the executed queries in the Local Storage of the browser, in the query settings, you can configure how many queries Studio will keep in history. You can also search a previously executed query, delete all the queries from the history or delete a single query.
Starting from Studio v.2.0, you can send the result set of a query to the Graph Editor by clicking on the circle icon in the result set actions. This allows you to visualize your data graphically.
JSON Output
Studio communicates with the OrientDB Server using HTTP/RESt+JSON protocol. To see the output in JSON format, press the RAW tab.
Query Explain
search: keywords: ['Studio', 'edit document']
Edit Document
search: keywords: ['Studio', 'edit vertex']
Edit Vertex
search: keywords: ['Studio', 'graph', 'Graph Editor']
Graph Editor
Since Studio 2.0 we have a new brand graph editor. Not only you can visualize your data in a graph way but you can also interact with the graph and modify it.
To populate the graph area just type a query in the query editor or use the functionality Send To Graph from the Browse UI
Supported operations in the Graph Editor are:
- Add Vertices
- Save the Graph Rendering Configuration
- Clear the Graph Rendering Canvas
- Delete Vertices
- Remove Vertices from Canvas
- Edit Vertices
- Inspect Vertices
- Change the Rendering Configuration of Vertices
- Navigating Relationships
- Create Edges between Vertices
- Delete Edges between Vertices
- Inspect Edges
- Edit Edges
Add Vertices
To add a new Vertex in your Graph Database and in the Graph Canvas area you have to press the button Add Vertex. This operation is done in two steps.
The first step you have to choose the class for the new Vertex and then click Next
In the second step you have to insert the fields values of the new vertex, you can also add custom fields as OrientDB supports Schema-Less mode. To make the new vertex persistent click to Save changes and the vertex will be saved into the database and added to the canvas area
Delete Vertices
Open the circular menu by clicking on the Vertex that you want to delete, open the sub-menu by passing hover the mouse to the menu entry more (...) and then click the trash icon.
Remove Vertices from Canvas
Open the circular menu , open the sub-menu by passing hover the mouse to the menu entry more (...) and then click the eraser icon.
Edit Vertices
Open the circular menu and then click to the edit icon, Studio will open a popup where you can edit the vertex properties.
Inspect Vertices
If you want to take a quick look to the Vertex property, click to the eye icon.
Change the Rendering Configuration of Vertices
Navigating Relationships
Create Edges between Vertices
Delete Edges between Vertices
Inspect Edges
Edit Edges
link to the schema manager
search: keywords: ['Studio', 'schema', 'schema manager']
Schema Manager Panel
OrientDB can work in schema-less mode, schema mode or a mix of both. Here we'll discuss the schema mode. To know more about schema in OrientDB go here
Here you can :
- Browse all the Classes of your database
- Create a new Class
- Rename/Drop a Class
- Change the cluster selection for a Class
- Edit a class by clicking on a class row in the table
- View all indexes created
View all indexes
When you want to have an overview of all indexes created in your database, just click the All indexes button in the Schema UI. This will provide quick access to some information about indexes (name, type, properties, etc) and you can drop or rebuild them from here.
search: keywords: ['Studio', 'class']
Classes
Creating a Class
To create a new Class, just click the New Class button. Some information is required to create the new class.
- Name
- SuperClass
- Alias (Optional)
- Abstract
Here you can find more information about Classes
Editing a Class
Dropping a Class
search: keywords: ['Studio', 'property']
Properties
Creating a Property
Editing a Property
Dropping a Property
search: keywords: ['Studio', 'index']
Indexes
Creating an Index
Rebuilding an Index
Dropping an Index
search: keywords: ['Studio', 'functions']
Functions Management Panel
OrientDB allows to extend the SQL language by providing Functions. Functions can be used also to create data-driven micro services. For more information look at Functions.
search: keywords: ['Studio', 'security']
Security Manager Panel
The Security Manager panel allows you to manage Database Users and Roles in a graphical way. For detailed information about Security in OrientDB, visit here
Tab Users
Here you can manage the database users:
- Search Users
- Add Users
- Delete Users
- Edit User: roles can be edited in-line, for name, status and password click the Edit button
Tab Roles
Here you can manage the database roles:
- Search Role
- Add Role
- Delete Role
- Edit Role
Users Management
Creating Database Users
To add a new User, click the Add User button, complete the information for the new user (name, password, status, roles) and then save to add the new user to the database.
Editing Database Users
Dropping Database Users
Roles Management
Creating Roles
To add a new User, click the Add Role button, complete the information for the new role (name, parent role, mode) and then save to add the new role to the database.
Deleting Roles
Adding Rules to a Role
To add a new security rule for the selected role, click the Add Rule button. This will ask you the string of the resource that you want to secure. For a list of available resources, see Resources
Then you can configure the CRUD permissions on the newly created resource.
Server-Level Commands
(since v 3.2)
Server-Level Commands allow SQL-like script execution on the server.
Since the very first versions, OrientDB allowed to execute a few different types of scripts, like SQL commands or console scripts.
SQL commands have a generic API that could be executed from any endpoint (REST, Studio, Console, application API) with the exact same effect. The scope of a SQL command is only a single database schema, so operations that do not involve the DB but have a broader scope (eg. the server instance or the cluster) can hardly be implemented with this mechanism.
Console scripts (eg. commands like info
or like the old create database
) act at server level, but they can only be executed via CLI. This makes this mechanism a bit limited and hard to extend.
In V 3.2 OrientDB introduces Server-Level Commands as a first-class component.
A Server-Level Command is a text command that is executed in the scope of a server, so the goal of this component is to run operations that do not limit their scope to a single database.
At this stage, the following commands are supported
How to invoke a Server-Level Command
Console
The console natively supports Server-Level Commands.
To run a command you have to be connected to a Server with connect env
eg. to create a MEMORY database using a Server-Level Command:
orientdb> connect env embedded:./target/dir root root
orientdb> create database test memory users (admin identified by 'admin' role admin)
Studio
From SERVER MANAGEMENT
Java API
Using OrientDB class, with the new execute()
methods
orientDB = new OrientDB("remote:localhost", "root", "root", OrientDBConfig.defaultConfig());
orientDB.execute(
"create database ? memory users (admin identified by 'admin' role admin)",
theDbName);
search: keywords: ['SQL', 'command', 'create', 'database', 'CREATE DATABASE']
Server Commands - CREATE DATABASE
Creates a database on the current environment/server
Syntax
CREATE DATABASE foo plocal users (foo identified by 'pippo' role admin, reader identified by ? role [reader, writer])
CREATE DATABASE <dbName> <type> [ USERS (<username> IDENTIFIED BY <password> ROLE <roleName>)*] [<configJson>]
<dbName>
The database name<type>
plocal
ormemory
<username>
the name of a user to add to the database<password>
the password of a user to add to the database<roleName>
the name of a role to assign to the user added to the database<configJson>
the custom configuration for current db
Examples
-
Create a DB:
orientdb>
CREATE DATABASE foo plocal
-
Create a DB with custom users:
orientdb>
CREATE DATABASE foo plocal users (foo identified by 'pippo' role admin, reader identified by ? role [reader, writer])
-
Create a DB with legacy default users:
orientdb>
CREATE DATABASE foo plocal {"config":{"security.createDefaultUsers": true}}
search: keywords: ['SQL', 'command', 'alter', 'class', 'ALTER CLASS']
Server Commands - DROP DATABASE
Drops a database
Syntax
DROP DATABASE <dbName> [if exists]
<dbName>
the name of an existing db
Examples
-
drop an existing db:
orientdb>
DROP DATABASE foo
-
drop a db if it exists:
orientdb>
DROP DATABASE foo if exists
search: keywords: ['SQL', 'command', 'alter', 'class', 'ALTER CLASS']
Server Commands - CREATE SYSTEM USER
Creates a System user
Syntax
CREATE SYSTEM USER <userName> IDENTIFIED BY <password> ROLE <role>
<userName>
the user name<password>
the password<role>
an existing role name
Examples
-
Create a system user with one role:
orientdb>
CREATE SYSTEM USER test IDENTIFIED BY 'foo' ROLE admin
-
Create a system user with multiple roles:
orientdb>
CREATE SYSTEM USER test IDENTIFIED BY 'foo' ROLE [reader, writer]
Server Commands - ALTER SYSTEM Role
Alters the properties of an existing system user
Syntax
ALTER SYSTEM ROLE <roleName> [SET POLICY <policyName ON <resource>]* [REMOVE POLICY ON <resource>]
<roleName>
A system role name<policyName>
The name of a security policy<resource>
A database resource, eg.database.class.AClassName
Examples
orientdb>
alter system role Foo set policy bar on database.class.Person
set policy bar on database.class.Xx remove policy on database.class.Person
Server Commands - EXISTS SYSTEM USER
Checks whether a system user exists
Syntax
EXISTS SYSTEM USER <userName>
<userName>
the system user name
Examples
orientdb> EXISTS SYSTEM USER test
Teleporter
OrientDB Teleporter is a tool that synchronizes a RDBMS to OrientDB database. You can use Teleporter to:
- Import your existing RDBMS to OrientDB
- Keep your OrientDB database synchronized with changes from the RDBMS. In this case the database on RDBMS remains the primary and the database on OrientDB a synchronized copy. Synchronization is one way, so all the changes in OrientDB database will not be propagated to the RDBMS
Teleporter is fully compatible with several RDBMS that have a JDBC driver: we successfully tested Teleporter with Oracle, SQLServer, MySQL, PostgreSQL and HyperSQL. Teleporter manages all the necessary type conversions between the different DBMSs and imports all your data as Graph in OrientDB.
NOTE: This feature is available both for the OrientDB Enterprise Edition and the OrientDB Community Edition. But beware: in community edition you can migrate your source relational database but you cannot enjoy the synchronize feature, only available in the enterprise edition.
How Teleporter works
Teleporter looks for the specific DBMS meta-data in order to perform a logical inference of the source DB schema for the building of a corresponding graph model. Eventually the data importing phase is performed.
Teleporter has a pluggable importing strategy. Two strategies are provided out of the box:
- naive strategy, the simplest one
- naive-aggregate strategy. It performs a "naive" import of the data source. The data source schema is translated semi-directly in a correspondent and coherent graph model using an aggregation policy on the junction tables of dimension equals to 2
To learn more about the two different execution strategies click here.
Usage
Teleporter is a tool written in Java, but can be used as a tool thanks to the teleporter.sh script (or .bat on Windows).
./oteleporter.sh -jdriver <jdbc-driver> -jurl <jdbc-url> -juser <username>
-jpasswd <password> -ourl <orientdb-url> [-s <strategy>]
[-nr <name-resolver>] [-v <verbose-level>]
([-include <table-names>] | [-exclude <table-names>])
[-inheritance <orm-technology>:<ORM-file-url>]
[-conf <configuration-file-location>]
Arguments
- -jdriver is the driver name of the DBMS from which you want to execute the import (it's not case sensitive)
- -jurl is the JDBC URL giving the location of the source database to import
- -ourl is the URL for the destination OrientDB graph database
- -juser (optional) is the username to access the source database
- -jpasswd (optional) is the password to access the source database
- -s (optional) is the strategy adopted during the importing phase. If not specified naive-aggregate strategy is adopted. Possible values:
- naive: performs a "naive" import of the data source. The data source schema is translated semi-directly in a correspondent and coherent graph model
- naive-aggregate: performs a "naive" import of the data source. The data source schema is translated semi-directly in a correspondent and coherent graph model using an aggregation policy on the junction tables of dimension equals to 2
- -nr (optional) is the name of the resolver which transforms the names of