Fork of https://github.com/oxigraph/oxigraph.git for the purpose of NextGraph project
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
3015 lines
109 KiB
3015 lines
109 KiB
#![allow(clippy::print_stderr, clippy::cast_precision_loss, clippy::use_debug)]
|
|
use anyhow::{bail, ensure, Context};
|
|
use clap::{Parser, Subcommand, ValueHint};
|
|
use flate2::read::MultiGzDecoder;
|
|
use oxhttp::model::{Body, HeaderName, HeaderValue, Method, Request, Response, Status};
|
|
use oxhttp::Server;
|
|
use oxigraph::io::{RdfFormat, RdfParser, RdfSerializer};
|
|
use oxigraph::model::{
|
|
GraphName, GraphNameRef, IriParseError, NamedNode, NamedNodeRef, NamedOrBlankNode,
|
|
};
|
|
use oxigraph::sparql::results::{QueryResultsFormat, QueryResultsSerializer};
|
|
use oxigraph::sparql::{Query, QueryOptions, QueryResults, Update};
|
|
use oxigraph::store::{BulkLoader, LoaderError, Store};
|
|
use oxiri::Iri;
|
|
use rand::random;
|
|
use rayon_core::ThreadPoolBuilder;
|
|
use std::borrow::Cow;
|
|
use std::cell::RefCell;
|
|
use std::cmp::{max, min};
|
|
#[cfg(target_os = "linux")]
|
|
use std::env;
|
|
use std::ffi::OsStr;
|
|
use std::fs::File;
|
|
use std::io::{self, stdin, stdout, BufWriter, Read, Write};
|
|
use std::net::ToSocketAddrs;
|
|
#[cfg(target_os = "linux")]
|
|
use std::os::unix::net::UnixDatagram;
|
|
use std::path::{Path, PathBuf};
|
|
use std::rc::Rc;
|
|
use std::str::FromStr;
|
|
use std::thread::available_parallelism;
|
|
use std::time::{Duration, Instant};
|
|
use std::{fmt, fs, str};
|
|
use url::form_urlencoded;
|
|
|
|
const MAX_SPARQL_BODY_SIZE: u64 = 0x0010_0000;
|
|
const HTTP_TIMEOUT: Duration = Duration::from_secs(60);
|
|
const HTML_ROOT_PAGE: &str = include_str!("../templates/query.html");
|
|
const LOGO: &str = include_str!("../logo.svg");
|
|
|
|
#[derive(Parser)]
|
|
#[command(about, version, name = "oxigraph")]
|
|
/// Oxigraph command line toolkit and SPARQL HTTP server.
|
|
struct Args {
|
|
#[command(subcommand)]
|
|
command: Command,
|
|
}
|
|
|
|
#[derive(Subcommand)]
|
|
enum Command {
|
|
/// Start Oxigraph HTTP server in read-write mode.
|
|
Serve {
|
|
/// Directory in which the data should be persisted.
|
|
///
|
|
/// If not present. An in-memory storage will be used.
|
|
#[arg(short, long, value_hint = ValueHint::DirPath)]
|
|
location: Option<PathBuf>,
|
|
/// Host and port to listen to.
|
|
#[arg(short, long, default_value = "localhost:7878", value_hint = ValueHint::Hostname)]
|
|
bind: String,
|
|
/// Allows cross-origin requests
|
|
#[arg(long)]
|
|
cors: bool,
|
|
},
|
|
/// Start Oxigraph HTTP server in read-only mode.
|
|
///
|
|
/// It allows to read the database while other processes are also reading it.
|
|
/// Opening as read-only while having an other process writing the database is undefined behavior.
|
|
/// Please use the serve-secondary command in this case.
|
|
ServeReadOnly {
|
|
/// Directory in which Oxigraph data are persisted.
|
|
#[arg(short, long, value_hint = ValueHint::DirPath)]
|
|
location: PathBuf,
|
|
/// Host and port to listen to.
|
|
#[arg(short, long, default_value = "localhost:7878")]
|
|
bind: String,
|
|
/// Allows cross-origin requests
|
|
#[arg(long)]
|
|
cors: bool,
|
|
},
|
|
/// Start Oxigraph HTTP server in secondary mode.
|
|
///
|
|
/// It allows to read the database while an other process is writing it.
|
|
/// Changes done while this process is running will be replicated after a possible lag.
|
|
///
|
|
/// Beware: RocksDB secondary mode does not support snapshots and transactions.
|
|
/// Dirty reads might happen.
|
|
ServeSecondary {
|
|
/// Directory where the primary Oxigraph instance is writing to.
|
|
#[arg(long, value_hint = ValueHint::DirPath)]
|
|
primary_location: PathBuf,
|
|
/// Directory to which the current secondary instance might write to.
|
|
///
|
|
/// By default, temporary storage is used.
|
|
#[arg(long, value_hint = ValueHint::DirPath)]
|
|
secondary_location: Option<PathBuf>,
|
|
/// Host and port to listen to.
|
|
#[arg(short, long, default_value = "localhost:7878")]
|
|
bind: String,
|
|
/// Allows cross-origin requests
|
|
#[arg(long)]
|
|
cors: bool,
|
|
},
|
|
/// Creates database backup into a target directory.
|
|
///
|
|
/// After its creation, the backup is usable a separated Oxigraph database
|
|
/// and operates independently from the original database.
|
|
///
|
|
/// If the target directory is in the same file system as the current database,
|
|
/// the database content will not be fully copied
|
|
/// but hard links will be used to point to the original database immutable snapshots.
|
|
/// This allows cheap regular backups.
|
|
///
|
|
/// If you want to move your data to another RDF storage system, you should use the dump operation instead.
|
|
Backup {
|
|
/// Directory in which Oxigraph data are persisted.
|
|
#[arg(short, long, value_hint = ValueHint::DirPath)]
|
|
location: PathBuf,
|
|
/// Directory in which the backup will be written.
|
|
#[arg(short, long, value_hint = ValueHint::DirPath)]
|
|
destination: PathBuf,
|
|
},
|
|
/// Load file(s) into the store.
|
|
///
|
|
/// Feel free to enable the --lenient option if you know your input is valid to get better performances.
|
|
Load {
|
|
/// Directory in which Oxigraph data are persisted.
|
|
#[arg(short, long, value_hint = ValueHint::DirPath)]
|
|
location: PathBuf,
|
|
/// File(s) to load.
|
|
///
|
|
/// If multiple files are provided they are loaded in parallel.
|
|
///
|
|
/// If no file is given, stdin is read.
|
|
#[arg(short, long, num_args = 0.., value_hint = ValueHint::FilePath)]
|
|
file: Vec<PathBuf>,
|
|
/// The format of the file(s) to load.
|
|
///
|
|
/// Can be an extension like "nt" or a MIME type like "application/n-triples".
|
|
///
|
|
/// By default the format is guessed from the loaded file extension.
|
|
#[arg(long, required_unless_present = "file")]
|
|
format: Option<String>,
|
|
/// Base IRI of the file(s) to load.
|
|
#[arg(long, value_hint = ValueHint::Url)]
|
|
base: Option<String>,
|
|
/// Attempt to keep loading even if the data file is invalid.
|
|
///
|
|
/// This disables most of validation on RDF content.
|
|
#[arg(long)]
|
|
lenient: bool,
|
|
/// Name of the graph to load the data to.
|
|
///
|
|
/// By default the default graph is used.
|
|
///
|
|
/// Only available when loading a graph file (N-Triples, Turtle...) and not a dataset file (N-Quads, TriG...).
|
|
#[arg(long, value_hint = ValueHint::Url)]
|
|
graph: Option<String>,
|
|
},
|
|
/// Dump the store content into a file.
|
|
Dump {
|
|
/// Directory in which Oxigraph data are persisted.
|
|
#[arg(short, long, value_hint = ValueHint::DirPath)]
|
|
location: PathBuf,
|
|
/// File to dump to.
|
|
///
|
|
/// If no file is given, stdout is used.
|
|
#[arg(short, long, value_hint = ValueHint::FilePath)]
|
|
file: Option<PathBuf>,
|
|
/// The format of the file(s) to dump.
|
|
///
|
|
/// Can be an extension like "nt" or a MIME type like "application/n-triples".
|
|
///
|
|
/// By default the format is guessed from the target file extension.
|
|
#[arg(long, required_unless_present = "file")]
|
|
format: Option<String>,
|
|
/// Name of the graph to dump.
|
|
///
|
|
/// By default all graphs are dumped if the output format supports datasets.
|
|
#[arg(long, value_hint = ValueHint::Url)]
|
|
graph: Option<String>,
|
|
},
|
|
/// Executes a SPARQL query against the store.
|
|
Query {
|
|
/// Directory in which Oxigraph data are persisted.
|
|
#[arg(short, long, value_hint = ValueHint::DirPath)]
|
|
location: PathBuf,
|
|
/// The SPARQL query to execute.
|
|
///
|
|
/// If no query or query file are given, stdin is used.
|
|
#[arg(short, long, conflicts_with = "query_file")]
|
|
query: Option<String>,
|
|
/// File in which the query is stored.
|
|
///
|
|
/// If no query or query file are given, stdin is used.
|
|
#[arg(long, conflicts_with = "query", value_hint = ValueHint::FilePath)]
|
|
query_file: Option<PathBuf>,
|
|
/// Base IRI of the query.
|
|
#[arg(long, value_hint = ValueHint::Url)]
|
|
query_base: Option<String>,
|
|
/// File in which the query results will be stored.
|
|
///
|
|
/// If no file is given, stdout is used.
|
|
#[arg(short, long, value_hint = ValueHint::FilePath)]
|
|
results_file: Option<PathBuf>,
|
|
/// The format of the results.
|
|
///
|
|
/// Can be an extension like "nt" or a MIME type like "application/n-triples".
|
|
///
|
|
/// By default the format is guessed from the results file extension.
|
|
#[arg(long, required_unless_present = "results_file")]
|
|
results_format: Option<String>,
|
|
/// Prints to stderr a human-readable explanation of the query evaluation.
|
|
///
|
|
/// Use the stats option to print also query evaluation statistics.
|
|
#[arg(long, conflicts_with = "explain_file")]
|
|
explain: bool,
|
|
/// Write to the given file an explanation of the query evaluation.
|
|
///
|
|
/// If the file extension is .json the JSON format is used, if .txt a human readable format is used.
|
|
///
|
|
/// Use the stats option to print also query evaluation statistics.
|
|
#[arg(long, conflicts_with = "explain", value_hint = ValueHint::FilePath)]
|
|
explain_file: Option<PathBuf>,
|
|
/// Computes some evaluation statistics to print as part of the query explanations.
|
|
///
|
|
/// Beware, computing the statistics adds some overhead to the evaluation runtime.
|
|
#[arg(long)]
|
|
stats: bool,
|
|
},
|
|
/// Executes a SPARQL update against the store.
|
|
Update {
|
|
/// Directory in which Oxigraph data are persisted.
|
|
#[arg(short, long, value_hint = ValueHint::DirPath)]
|
|
location: PathBuf,
|
|
/// The SPARQL update to execute.
|
|
///
|
|
/// If no query or query file are given, stdin is used.
|
|
#[arg(short, long, conflicts_with = "update_file")]
|
|
update: Option<String>,
|
|
/// File in which the update is stored.
|
|
///
|
|
/// If no update or update file are given, stdin is used.
|
|
#[arg(long, conflicts_with = "update", value_hint = ValueHint::FilePath)]
|
|
update_file: Option<PathBuf>,
|
|
/// Base IRI of the update.
|
|
#[arg(long, value_hint = ValueHint::Url)]
|
|
update_base: Option<String>,
|
|
},
|
|
/// Optimizes the database storage.
|
|
///
|
|
/// Done by default in the background when serving requests.
|
|
/// It is likely to not be useful in most of cases except if you provide a read-only SPARQL endpoint under heavy load.
|
|
Optimize {
|
|
/// Directory in which Oxigraph data are persisted.
|
|
#[arg(short, long, value_hint = ValueHint::DirPath)]
|
|
location: PathBuf,
|
|
},
|
|
/// Converts a RDF serialization from one format to an other.
|
|
Convert {
|
|
/// File to convert from.
|
|
///
|
|
/// If no file is given, stdin is read.
|
|
#[arg(short, long, value_hint = ValueHint::FilePath)]
|
|
from_file: Option<PathBuf>,
|
|
/// The format of the file(s) to convert from.
|
|
///
|
|
/// Can be an extension like "nt" or a MIME type like "application/n-triples".
|
|
///
|
|
/// By default the format is guessed from the input file extension.
|
|
#[arg(long, required_unless_present = "from_file")]
|
|
from_format: Option<String>,
|
|
/// Base IRI of the file to read.
|
|
#[arg(long, value_hint = ValueHint::Url)]
|
|
from_base: Option<String>,
|
|
/// File to convert to.
|
|
///
|
|
/// If no file is given, stdout is written.
|
|
#[arg(short, long, value_hint = ValueHint::FilePath)]
|
|
to_file: Option<PathBuf>,
|
|
/// The format of the file(s) to convert from.
|
|
///
|
|
/// Can be an extension like "nt" or a MIME type like "application/n-triples".
|
|
///
|
|
/// By default the format is guessed from the target file extension.
|
|
#[arg(long, required_unless_present = "to_file")]
|
|
to_format: Option<String>,
|
|
/// Attempt to keep converting even if the data file is invalid.
|
|
#[arg(long)]
|
|
lenient: bool,
|
|
/// Only load the given named graph from the input file.
|
|
///
|
|
/// By default all graphs are loaded.
|
|
#[arg(long, conflicts_with = "from_default_graph", value_hint = ValueHint::Url)]
|
|
from_graph: Option<String>,
|
|
/// Only load the default graph from the input file.
|
|
#[arg(long, conflicts_with = "from_graph")]
|
|
from_default_graph: bool,
|
|
/// Name of the graph to map the default graph to.
|
|
///
|
|
/// By default the default graph is used.
|
|
#[arg(long, value_hint = ValueHint::Url)]
|
|
to_graph: Option<String>,
|
|
},
|
|
}
|
|
|
|
pub fn main() -> anyhow::Result<()> {
|
|
let matches = Args::parse();
|
|
match matches.command {
|
|
Command::Serve {
|
|
location,
|
|
bind,
|
|
cors,
|
|
} => serve(
|
|
if let Some(location) = location {
|
|
Store::open(location)
|
|
} else {
|
|
Store::new()
|
|
}?,
|
|
&bind,
|
|
false,
|
|
cors,
|
|
),
|
|
Command::ServeReadOnly {
|
|
location,
|
|
bind,
|
|
cors,
|
|
} => serve(Store::open_read_only(location)?, &bind, true, cors),
|
|
Command::ServeSecondary {
|
|
primary_location,
|
|
secondary_location,
|
|
bind,
|
|
cors,
|
|
} => serve(
|
|
if let Some(secondary_location) = secondary_location {
|
|
Store::open_persistent_secondary(primary_location, secondary_location)
|
|
} else {
|
|
Store::open_secondary(primary_location)
|
|
}?,
|
|
&bind,
|
|
true,
|
|
cors,
|
|
),
|
|
Command::Backup {
|
|
location,
|
|
destination,
|
|
} => {
|
|
let store = Store::open_read_only(location)?;
|
|
store.backup(destination)?;
|
|
Ok(())
|
|
}
|
|
Command::Load {
|
|
location,
|
|
file,
|
|
lenient,
|
|
format,
|
|
base,
|
|
graph,
|
|
} => {
|
|
let store = Store::open(location)?;
|
|
let format = if let Some(format) = format {
|
|
Some(rdf_format_from_name(&format)?)
|
|
} else {
|
|
None
|
|
};
|
|
let graph = if let Some(iri) = &graph {
|
|
Some(
|
|
NamedNode::new(iri)
|
|
.with_context(|| format!("The target graph name {iri} is invalid"))?,
|
|
)
|
|
} else {
|
|
None
|
|
};
|
|
#[allow(clippy::cast_precision_loss)]
|
|
if file.is_empty() {
|
|
// We read from stdin
|
|
let start = Instant::now();
|
|
let mut loader = store.bulk_loader().on_progress(move |size| {
|
|
let elapsed = start.elapsed();
|
|
eprintln!(
|
|
"{size} triples loaded in {}s ({} t/s)",
|
|
elapsed.as_secs(),
|
|
((size as f64) / elapsed.as_secs_f64()).round()
|
|
)
|
|
});
|
|
if lenient {
|
|
loader = loader.on_parse_error(move |e| {
|
|
eprintln!("Parsing error: {e}");
|
|
Ok(())
|
|
})
|
|
}
|
|
bulk_load(
|
|
&loader,
|
|
stdin().lock(),
|
|
format.context("The --format option must be set when loading from stdin")?,
|
|
base.as_deref(),
|
|
graph,
|
|
lenient,
|
|
)
|
|
} else {
|
|
ThreadPoolBuilder::new()
|
|
.num_threads(max(1, available_parallelism()?.get() / 2))
|
|
.thread_name(|i| format!("Oxigraph bulk loader thread {i}"))
|
|
.build()?
|
|
.scope(|s| {
|
|
for file in file {
|
|
let store = store.clone();
|
|
let graph = graph.clone();
|
|
let base = base.clone();
|
|
s.spawn(move |_| {
|
|
let f = file.clone();
|
|
let start = Instant::now();
|
|
let mut loader = store.bulk_loader().on_progress(move |size| {
|
|
let elapsed = start.elapsed();
|
|
eprintln!(
|
|
"{} triples loaded in {}s ({} t/s) from {}",
|
|
size,
|
|
elapsed.as_secs(),
|
|
((size as f64) / elapsed.as_secs_f64()).round(),
|
|
f.display()
|
|
)
|
|
});
|
|
if lenient {
|
|
let f = file.clone();
|
|
loader = loader.on_parse_error(move |e| {
|
|
eprintln!("Parsing error on file {}: {}", f.display(), e);
|
|
Ok(())
|
|
})
|
|
}
|
|
let fp = match File::open(&file) {
|
|
Ok(fp) => fp,
|
|
Err(error) => {
|
|
eprintln!(
|
|
"Error while opening file {}: {}",
|
|
file.display(),
|
|
error
|
|
);
|
|
return;
|
|
}
|
|
};
|
|
if let Err(error) = {
|
|
if file.extension().map_or(false, |e| e == OsStr::new("gz")) {
|
|
bulk_load(
|
|
&loader,
|
|
MultiGzDecoder::new(fp),
|
|
format.unwrap_or_else(|| {
|
|
rdf_format_from_path(&file.with_extension(""))
|
|
.unwrap()
|
|
}),
|
|
base.as_deref(),
|
|
graph,
|
|
lenient,
|
|
)
|
|
} else {
|
|
bulk_load(
|
|
&loader,
|
|
fp,
|
|
format.unwrap_or_else(|| {
|
|
rdf_format_from_path(&file).unwrap()
|
|
}),
|
|
base.as_deref(),
|
|
graph,
|
|
lenient,
|
|
)
|
|
}
|
|
} {
|
|
eprintln!(
|
|
"Error while loading file {}: {}",
|
|
file.display(),
|
|
error
|
|
)
|
|
// TODO: hard fail
|
|
}
|
|
})
|
|
}
|
|
});
|
|
store.flush()?;
|
|
Ok(())
|
|
}
|
|
}
|
|
Command::Dump {
|
|
location,
|
|
file,
|
|
format,
|
|
graph,
|
|
} => {
|
|
let store = Store::open_read_only(location)?;
|
|
let format = if let Some(format) = format {
|
|
rdf_format_from_name(&format)?
|
|
} else if let Some(file) = &file {
|
|
rdf_format_from_path(file)?
|
|
} else {
|
|
bail!("The --format option must be set when writing to stdout")
|
|
};
|
|
let graph = if let Some(graph) = &graph {
|
|
Some(if graph.eq_ignore_ascii_case("default") {
|
|
GraphNameRef::DefaultGraph
|
|
} else {
|
|
NamedNodeRef::new(graph)
|
|
.with_context(|| format!("The target graph name {graph} is invalid"))?
|
|
.into()
|
|
})
|
|
} else {
|
|
None
|
|
};
|
|
if let Some(file) = file {
|
|
close_file_writer(dump(
|
|
&store,
|
|
BufWriter::new(File::create(file)?),
|
|
format,
|
|
graph,
|
|
)?)?;
|
|
} else {
|
|
dump(&store, stdout().lock(), format, graph)?.flush()?;
|
|
}
|
|
Ok(())
|
|
}
|
|
Command::Query {
|
|
location,
|
|
query,
|
|
query_file,
|
|
query_base,
|
|
results_file,
|
|
results_format,
|
|
explain,
|
|
explain_file,
|
|
stats,
|
|
} => {
|
|
let query = if let Some(query) = query {
|
|
query
|
|
} else if let Some(query_file) = query_file {
|
|
fs::read_to_string(&query_file).with_context(|| {
|
|
format!("Not able to read query file {}", query_file.display())
|
|
})?
|
|
} else {
|
|
io::read_to_string(stdin().lock())?
|
|
};
|
|
let query = Query::parse(&query, query_base.as_deref())?;
|
|
let store = Store::open_read_only(location)?;
|
|
let (results, explanation) =
|
|
store.explain_query_opt(query, QueryOptions::default(), stats)?;
|
|
let print_result = (|| {
|
|
match results? {
|
|
QueryResults::Solutions(solutions) => {
|
|
let format = if let Some(name) = results_format {
|
|
if let Some(format) = QueryResultsFormat::from_extension(&name) {
|
|
format
|
|
} else if let Some(format) = QueryResultsFormat::from_media_type(&name)
|
|
{
|
|
format
|
|
} else {
|
|
bail!("The file format '{name}' is unknown")
|
|
}
|
|
} else if let Some(results_file) = &results_file {
|
|
format_from_path(results_file, |ext| {
|
|
QueryResultsFormat::from_extension(ext).with_context(|| {
|
|
format!("The file extension '{ext}' is unknown")
|
|
})
|
|
})?
|
|
} else {
|
|
bail!("The --results-format option must be set when writing to stdout")
|
|
};
|
|
if let Some(results_file) = results_file {
|
|
let mut writer = QueryResultsSerializer::from_format(format)
|
|
.serialize_solutions_to_write(
|
|
BufWriter::new(File::create(results_file)?),
|
|
solutions.variables().to_vec(),
|
|
)?;
|
|
for solution in solutions {
|
|
writer.write(&solution?)?;
|
|
}
|
|
close_file_writer(writer.finish()?)?;
|
|
} else {
|
|
let mut writer = QueryResultsSerializer::from_format(format)
|
|
.serialize_solutions_to_write(
|
|
stdout().lock(),
|
|
solutions.variables().to_vec(),
|
|
)?;
|
|
for solution in solutions {
|
|
writer.write(&solution?)?;
|
|
}
|
|
writer.finish()?.flush()?;
|
|
}
|
|
}
|
|
QueryResults::Boolean(result) => {
|
|
let format = if let Some(name) = results_format {
|
|
if let Some(format) = QueryResultsFormat::from_extension(&name) {
|
|
format
|
|
} else if let Some(format) = QueryResultsFormat::from_media_type(&name)
|
|
{
|
|
format
|
|
} else {
|
|
bail!("The file format '{name}' is unknown")
|
|
}
|
|
} else if let Some(results_file) = &results_file {
|
|
format_from_path(results_file, |ext| {
|
|
QueryResultsFormat::from_extension(ext).with_context(|| {
|
|
format!("The file extension '{ext}' is unknown")
|
|
})
|
|
})?
|
|
} else {
|
|
bail!("The --results-format option must be set when writing to stdout")
|
|
};
|
|
if let Some(results_file) = results_file {
|
|
close_file_writer(
|
|
QueryResultsSerializer::from_format(format)
|
|
.serialize_boolean_to_write(
|
|
BufWriter::new(File::create(results_file)?),
|
|
result,
|
|
)?,
|
|
)?;
|
|
} else {
|
|
QueryResultsSerializer::from_format(format)
|
|
.serialize_boolean_to_write(stdout().lock(), result)?
|
|
.flush()?;
|
|
}
|
|
}
|
|
QueryResults::Graph(triples) => {
|
|
let format = if let Some(name) = &results_format {
|
|
rdf_format_from_name(name)
|
|
} else if let Some(results_file) = &results_file {
|
|
rdf_format_from_path(results_file)
|
|
} else {
|
|
bail!("The --results-format option must be set when writing to stdout")
|
|
}?;
|
|
let serializer = RdfSerializer::from_format(format);
|
|
if let Some(results_file) = results_file {
|
|
let mut writer = serializer
|
|
.serialize_to_write(BufWriter::new(File::create(results_file)?));
|
|
for triple in triples {
|
|
writer.write_triple(triple?.as_ref())?;
|
|
}
|
|
close_file_writer(writer.finish()?)?;
|
|
} else {
|
|
let mut writer = serializer.serialize_to_write(stdout().lock());
|
|
for triple in triples {
|
|
writer.write_triple(triple?.as_ref())?;
|
|
}
|
|
writer.finish()?.flush()?;
|
|
}
|
|
}
|
|
}
|
|
Ok(())
|
|
})();
|
|
if let Some(explain_file) = explain_file {
|
|
let mut file = BufWriter::new(File::create(&explain_file)?);
|
|
match explain_file
|
|
.extension()
|
|
.and_then(OsStr::to_str) {
|
|
Some("json") => {
|
|
explanation.write_in_json(&mut file)?;
|
|
},
|
|
Some("txt") => {
|
|
write!(file, "{explanation:?}")?;
|
|
},
|
|
_ => bail!("The given explanation file {} must have an extension that is .json or .txt", explain_file.display())
|
|
}
|
|
close_file_writer(file)?;
|
|
} else if explain || stats {
|
|
eprintln!("{explanation:#?}");
|
|
}
|
|
print_result
|
|
}
|
|
Command::Update {
|
|
location,
|
|
update,
|
|
update_file,
|
|
update_base,
|
|
} => {
|
|
let update = if let Some(update) = update {
|
|
update
|
|
} else if let Some(update_file) = update_file {
|
|
fs::read_to_string(&update_file).with_context(|| {
|
|
format!("Not able to read update file {}", update_file.display())
|
|
})?
|
|
} else {
|
|
io::read_to_string(stdin().lock())?
|
|
};
|
|
let update = Update::parse(&update, update_base.as_deref())?;
|
|
let store = Store::open(location)?;
|
|
store.update(update)?;
|
|
store.flush()?;
|
|
Ok(())
|
|
}
|
|
Command::Optimize { location } => {
|
|
let store = Store::open(location)?;
|
|
store.optimize()?;
|
|
Ok(())
|
|
}
|
|
Command::Convert {
|
|
from_file,
|
|
from_format,
|
|
from_base,
|
|
to_file,
|
|
to_format,
|
|
lenient,
|
|
from_graph,
|
|
from_default_graph,
|
|
to_graph,
|
|
} => {
|
|
let from_format = if let Some(format) = from_format {
|
|
rdf_format_from_name(&format)?
|
|
} else if let Some(file) = &from_file {
|
|
rdf_format_from_path(file)?
|
|
} else {
|
|
bail!("The --from-format option must be set when reading from stdin")
|
|
};
|
|
let mut parser = RdfParser::from_format(from_format);
|
|
if let Some(base) = from_base {
|
|
parser = parser
|
|
.with_base_iri(&base)
|
|
.with_context(|| format!("Invalid base IRI {base}"))?;
|
|
}
|
|
|
|
let to_format = if let Some(format) = to_format {
|
|
rdf_format_from_name(&format)?
|
|
} else if let Some(file) = &to_file {
|
|
rdf_format_from_path(file)?
|
|
} else {
|
|
bail!("The --to-format option must be set when writing to stdout")
|
|
};
|
|
let serializer = RdfSerializer::from_format(to_format);
|
|
|
|
let from_graph = if let Some(from_graph) = from_graph {
|
|
Some(
|
|
NamedNode::new(&from_graph)
|
|
.with_context(|| format!("The source graph name {from_graph} is invalid"))?
|
|
.into(),
|
|
)
|
|
} else if from_default_graph {
|
|
Some(GraphName::DefaultGraph)
|
|
} else {
|
|
None
|
|
};
|
|
let to_graph = if let Some(to_graph) = to_graph {
|
|
NamedNode::new(&to_graph)
|
|
.with_context(|| format!("The target graph name {to_graph} is invalid"))?
|
|
.into()
|
|
} else {
|
|
GraphName::DefaultGraph
|
|
};
|
|
|
|
match (from_file, to_file) {
|
|
(Some(from_file), Some(to_file)) => close_file_writer(do_convert(
|
|
parser,
|
|
File::open(from_file)?,
|
|
serializer,
|
|
BufWriter::new(File::create(to_file)?),
|
|
lenient,
|
|
&from_graph,
|
|
&to_graph,
|
|
)?),
|
|
(Some(from_file), None) => do_convert(
|
|
parser,
|
|
File::open(from_file)?,
|
|
serializer,
|
|
stdout().lock(),
|
|
lenient,
|
|
&from_graph,
|
|
&to_graph,
|
|
)?
|
|
.flush(),
|
|
(None, Some(to_file)) => close_file_writer(do_convert(
|
|
parser,
|
|
stdin().lock(),
|
|
serializer,
|
|
BufWriter::new(File::create(to_file)?),
|
|
lenient,
|
|
&from_graph,
|
|
&to_graph,
|
|
)?),
|
|
(None, None) => do_convert(
|
|
parser,
|
|
stdin().lock(),
|
|
serializer,
|
|
stdout().lock(),
|
|
lenient,
|
|
&from_graph,
|
|
&to_graph,
|
|
)?
|
|
.flush(),
|
|
}?;
|
|
Ok(())
|
|
}
|
|
}
|
|
}
|
|
|
|
fn bulk_load(
|
|
loader: &BulkLoader,
|
|
read: impl Read,
|
|
format: RdfFormat,
|
|
base_iri: Option<&str>,
|
|
to_graph_name: Option<NamedNode>,
|
|
lenient: bool,
|
|
) -> anyhow::Result<()> {
|
|
let mut parser = RdfParser::from_format(format);
|
|
if let Some(to_graph_name) = to_graph_name {
|
|
parser = parser.with_default_graph(to_graph_name);
|
|
}
|
|
if let Some(base_iri) = base_iri {
|
|
parser = parser
|
|
.with_base_iri(base_iri)
|
|
.with_context(|| format!("Invalid base IRI {base_iri}"))?;
|
|
}
|
|
if lenient {
|
|
parser = parser.unchecked();
|
|
}
|
|
loader.load_from_read(parser, read)?;
|
|
Ok(())
|
|
}
|
|
|
|
fn dump<W: Write>(
|
|
store: &Store,
|
|
write: W,
|
|
format: RdfFormat,
|
|
from_graph_name: Option<GraphNameRef<'_>>,
|
|
) -> anyhow::Result<W> {
|
|
ensure!(format.supports_datasets() || from_graph_name.is_some(), "The --graph option is required when writing a format not supporting datasets like NTriples, Turtle or RDF/XML");
|
|
Ok(if let Some(from_graph_name) = from_graph_name {
|
|
store.dump_graph_to_write(from_graph_name, format, write)
|
|
} else {
|
|
store.dump_to_write(format, write)
|
|
}?)
|
|
}
|
|
|
|
fn do_convert<R: Read, W: Write>(
|
|
parser: RdfParser,
|
|
read: R,
|
|
mut serializer: RdfSerializer,
|
|
write: W,
|
|
lenient: bool,
|
|
from_graph: &Option<GraphName>,
|
|
default_graph: &GraphName,
|
|
) -> anyhow::Result<W> {
|
|
let mut parser = parser.parse_read(read);
|
|
let first = parser.next(); // We read the first element to get prefixes
|
|
for (prefix_name, prefix_iri) in parser.prefixes() {
|
|
serializer = serializer
|
|
.with_prefix(prefix_name, prefix_iri)
|
|
.with_context(|| format!("Invalid IRI for prefix {prefix_name}: {prefix_iri}"))?;
|
|
}
|
|
let mut writer = serializer.serialize_to_write(write);
|
|
for quad_result in first.into_iter().chain(parser) {
|
|
match quad_result {
|
|
Ok(mut quad) => {
|
|
if let Some(from_graph) = from_graph {
|
|
if quad.graph_name == *from_graph {
|
|
quad.graph_name = GraphName::DefaultGraph;
|
|
} else {
|
|
continue;
|
|
}
|
|
}
|
|
if quad.graph_name.is_default_graph() {
|
|
quad.graph_name = default_graph.clone();
|
|
}
|
|
writer.write_quad(&quad)?;
|
|
}
|
|
Err(e) => {
|
|
if lenient {
|
|
eprintln!("Parsing error: {e}");
|
|
} else {
|
|
return Err(e.into());
|
|
}
|
|
}
|
|
}
|
|
}
|
|
Ok(writer.finish()?)
|
|
}
|
|
|
|
fn format_from_path<T>(
|
|
path: &Path,
|
|
from_extension: impl FnOnce(&str) -> anyhow::Result<T>,
|
|
) -> anyhow::Result<T> {
|
|
if let Some(ext) = path.extension().and_then(OsStr::to_str) {
|
|
from_extension(ext).map_err(|e| {
|
|
e.context(format!(
|
|
"Not able to guess the file format from file name extension '{ext}'"
|
|
))
|
|
})
|
|
} else {
|
|
bail!(
|
|
"The path {} has no extension to guess a file format from",
|
|
path.display()
|
|
)
|
|
}
|
|
}
|
|
|
|
fn rdf_format_from_path(path: &Path) -> anyhow::Result<RdfFormat> {
|
|
format_from_path(path, |ext| {
|
|
RdfFormat::from_extension(ext)
|
|
.with_context(|| format!("The file extension '{ext}' is unknown"))
|
|
})
|
|
}
|
|
|
|
fn rdf_format_from_name(name: &str) -> anyhow::Result<RdfFormat> {
|
|
if let Some(t) = RdfFormat::from_extension(name) {
|
|
return Ok(t);
|
|
}
|
|
if let Some(t) = RdfFormat::from_media_type(name) {
|
|
return Ok(t);
|
|
}
|
|
bail!("The file format '{name}' is unknown")
|
|
}
|
|
|
|
fn serve(store: Store, bind: &str, read_only: bool, cors: bool) -> anyhow::Result<()> {
|
|
let mut server = if cors {
|
|
Server::new(cors_middleware(move |request| {
|
|
handle_request(request, store.clone(), read_only)
|
|
.unwrap_or_else(|(status, message)| error(status, message))
|
|
}))
|
|
} else {
|
|
Server::new(move |request| {
|
|
handle_request(request, store.clone(), read_only)
|
|
.unwrap_or_else(|(status, message)| error(status, message))
|
|
})
|
|
}
|
|
.with_global_timeout(HTTP_TIMEOUT)
|
|
.with_server_name(concat!("Oxigraph/", env!("CARGO_PKG_VERSION")))?
|
|
.with_max_concurrent_connections(available_parallelism()?.get() * 128);
|
|
for socket in bind.to_socket_addrs()? {
|
|
server = server.bind(socket);
|
|
}
|
|
let server = server.spawn()?;
|
|
#[cfg(target_os = "linux")]
|
|
systemd_notify_ready()?;
|
|
eprintln!("Listening for requests at http://{bind}");
|
|
server.join()?;
|
|
Ok(())
|
|
}
|
|
|
|
fn cors_middleware(
|
|
on_request: impl Fn(&mut Request) -> Response + Send + Sync + 'static,
|
|
) -> impl Fn(&mut Request) -> Response + Send + Sync + 'static {
|
|
let origin = HeaderName::from_str("Origin").unwrap();
|
|
let access_control_allow_origin = HeaderName::from_str("Access-Control-Allow-Origin").unwrap();
|
|
let access_control_request_method =
|
|
HeaderName::from_str("Access-Control-Request-Method").unwrap();
|
|
let access_control_allow_method = HeaderName::from_str("Access-Control-Allow-Methods").unwrap();
|
|
let access_control_request_headers =
|
|
HeaderName::from_str("Access-Control-Request-Headers").unwrap();
|
|
let access_control_allow_headers =
|
|
HeaderName::from_str("Access-Control-Allow-Headers").unwrap();
|
|
let star = HeaderValue::from_str("*").unwrap();
|
|
move |request| {
|
|
if *request.method() == Method::OPTIONS {
|
|
let mut response = Response::builder(Status::NO_CONTENT);
|
|
if request.header(&origin).is_some() {
|
|
response
|
|
.headers_mut()
|
|
.append(access_control_allow_origin.clone(), star.clone());
|
|
}
|
|
if let Some(method) = request.header(&access_control_request_method) {
|
|
response
|
|
.headers_mut()
|
|
.append(access_control_allow_method.clone(), method.clone());
|
|
}
|
|
if let Some(headers) = request.header(&access_control_request_headers) {
|
|
response
|
|
.headers_mut()
|
|
.append(access_control_allow_headers.clone(), headers.clone());
|
|
}
|
|
response.build()
|
|
} else {
|
|
let mut response = on_request(request);
|
|
if request.header(&origin).is_some() {
|
|
response
|
|
.headers_mut()
|
|
.append(access_control_allow_origin.clone(), star.clone());
|
|
}
|
|
response
|
|
}
|
|
}
|
|
}
|
|
|
|
type HttpError = (Status, String);
|
|
|
|
fn handle_request(
|
|
request: &mut Request,
|
|
store: Store,
|
|
read_only: bool,
|
|
) -> Result<Response, HttpError> {
|
|
match (request.url().path(), request.method().as_ref()) {
|
|
("/", "HEAD") => Ok(Response::builder(Status::OK)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text_html")
|
|
.unwrap()
|
|
.build()),
|
|
("/", "GET") => Ok(Response::builder(Status::OK)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text_html")
|
|
.unwrap()
|
|
.with_body(HTML_ROOT_PAGE)),
|
|
("/logo.svg", "HEAD") => Ok(Response::builder(Status::OK)
|
|
.with_header(HeaderName::CONTENT_TYPE, "image/svg+xml")
|
|
.unwrap()
|
|
.build()),
|
|
("/logo.svg", "GET") => Ok(Response::builder(Status::OK)
|
|
.with_header(HeaderName::CONTENT_TYPE, "image/svg+xml")
|
|
.unwrap()
|
|
.with_body(LOGO)),
|
|
("/query", "GET") => {
|
|
configure_and_evaluate_sparql_query(&store, &[url_query(request)], None, request)
|
|
}
|
|
("/query", "POST") => {
|
|
let content_type =
|
|
content_type(request).ok_or_else(|| bad_request("No Content-Type given"))?;
|
|
if content_type == "application/sparql-query" {
|
|
let query = io::read_to_string(request.body_mut().take(MAX_SPARQL_BODY_SIZE))
|
|
.map_err(bad_request)?;
|
|
configure_and_evaluate_sparql_query(
|
|
&store,
|
|
&[url_query(request)],
|
|
Some(query),
|
|
request,
|
|
)
|
|
} else if content_type == "application/x-www-form-urlencoded" {
|
|
let mut buffer = Vec::new();
|
|
request
|
|
.body_mut()
|
|
.take(MAX_SPARQL_BODY_SIZE)
|
|
.read_to_end(&mut buffer)
|
|
.map_err(bad_request)?;
|
|
configure_and_evaluate_sparql_query(
|
|
&store,
|
|
&[url_query(request), &buffer],
|
|
None,
|
|
request,
|
|
)
|
|
} else {
|
|
Err(unsupported_media_type(&content_type))
|
|
}
|
|
}
|
|
("/update", "POST") => {
|
|
if read_only {
|
|
return Err(the_server_is_read_only());
|
|
}
|
|
let content_type =
|
|
content_type(request).ok_or_else(|| bad_request("No Content-Type given"))?;
|
|
if content_type == "application/sparql-update" {
|
|
let update = io::read_to_string(request.body_mut().take(MAX_SPARQL_BODY_SIZE))
|
|
.map_err(bad_request)?;
|
|
configure_and_evaluate_sparql_update(
|
|
&store,
|
|
&[url_query(request)],
|
|
Some(update),
|
|
request,
|
|
)
|
|
} else if content_type == "application/x-www-form-urlencoded" {
|
|
let mut buffer = Vec::new();
|
|
request
|
|
.body_mut()
|
|
.take(MAX_SPARQL_BODY_SIZE)
|
|
.read_to_end(&mut buffer)
|
|
.map_err(bad_request)?;
|
|
configure_and_evaluate_sparql_update(
|
|
&store,
|
|
&[url_query(request), &buffer],
|
|
None,
|
|
request,
|
|
)
|
|
} else {
|
|
Err(unsupported_media_type(&content_type))
|
|
}
|
|
}
|
|
(path, "GET") if path.starts_with("/store") => {
|
|
if let Some(target) = store_target(request)? {
|
|
assert_that_graph_exists(&store, &target)?;
|
|
let format = rdf_content_negotiation(request)?;
|
|
|
|
let quads = store.quads_for_pattern(
|
|
None,
|
|
None,
|
|
None,
|
|
Some(GraphName::from(target).as_ref()),
|
|
);
|
|
ReadForWrite::build_response(
|
|
move |w| {
|
|
Ok((
|
|
RdfSerializer::from_format(format).serialize_to_write(w),
|
|
quads,
|
|
))
|
|
},
|
|
|(mut writer, mut quads)| {
|
|
Ok(if let Some(q) = quads.next() {
|
|
writer.write_triple(&q?.into())?;
|
|
Some((writer, quads))
|
|
} else {
|
|
writer.finish()?;
|
|
None
|
|
})
|
|
},
|
|
format.media_type(),
|
|
)
|
|
} else {
|
|
let format = rdf_content_negotiation(request)?;
|
|
if !format.supports_datasets() {
|
|
return Err(bad_request(format!(
|
|
"It is not possible to serialize the full RDF dataset using {format} that does not support named graphs"
|
|
)));
|
|
}
|
|
ReadForWrite::build_response(
|
|
move |w| {
|
|
Ok((
|
|
RdfSerializer::from_format(format).serialize_to_write(w),
|
|
store.iter(),
|
|
))
|
|
},
|
|
|(mut writer, mut quads)| {
|
|
Ok(if let Some(q) = quads.next() {
|
|
writer.write_quad(&q?)?;
|
|
Some((writer, quads))
|
|
} else {
|
|
writer.finish()?;
|
|
None
|
|
})
|
|
},
|
|
format.media_type(),
|
|
)
|
|
}
|
|
}
|
|
(path, "PUT") if path.starts_with("/store") => {
|
|
if read_only {
|
|
return Err(the_server_is_read_only());
|
|
}
|
|
let content_type =
|
|
content_type(request).ok_or_else(|| bad_request("No Content-Type given"))?;
|
|
if let Some(target) = store_target(request)? {
|
|
let format = RdfFormat::from_media_type(&content_type)
|
|
.ok_or_else(|| unsupported_media_type(&content_type))?;
|
|
let new = !match &target {
|
|
NamedGraphName::NamedNode(target) => {
|
|
if store
|
|
.contains_named_graph(target)
|
|
.map_err(internal_server_error)?
|
|
{
|
|
store.clear_graph(target).map_err(internal_server_error)?;
|
|
true
|
|
} else {
|
|
store
|
|
.insert_named_graph(target)
|
|
.map_err(internal_server_error)?;
|
|
false
|
|
}
|
|
}
|
|
NamedGraphName::DefaultGraph => {
|
|
store
|
|
.clear_graph(GraphNameRef::DefaultGraph)
|
|
.map_err(internal_server_error)?;
|
|
true
|
|
}
|
|
};
|
|
web_load_graph(&store, request, format, &GraphName::from(target))?;
|
|
Ok(Response::builder(if new {
|
|
Status::CREATED
|
|
} else {
|
|
Status::NO_CONTENT
|
|
})
|
|
.build())
|
|
} else {
|
|
let format = RdfFormat::from_media_type(&content_type)
|
|
.ok_or_else(|| unsupported_media_type(&content_type))?;
|
|
store.clear().map_err(internal_server_error)?;
|
|
web_load_dataset(&store, request, format)?;
|
|
Ok(Response::builder(Status::NO_CONTENT).build())
|
|
}
|
|
}
|
|
(path, "DELETE") if path.starts_with("/store") => {
|
|
if read_only {
|
|
return Err(the_server_is_read_only());
|
|
}
|
|
if let Some(target) = store_target(request)? {
|
|
match target {
|
|
NamedGraphName::DefaultGraph => store
|
|
.clear_graph(GraphNameRef::DefaultGraph)
|
|
.map_err(internal_server_error)?,
|
|
NamedGraphName::NamedNode(target) => {
|
|
if store
|
|
.contains_named_graph(&target)
|
|
.map_err(internal_server_error)?
|
|
{
|
|
store
|
|
.remove_named_graph(&target)
|
|
.map_err(internal_server_error)?;
|
|
} else {
|
|
return Err((
|
|
Status::NOT_FOUND,
|
|
format!("The graph {target} does not exists"),
|
|
));
|
|
}
|
|
}
|
|
}
|
|
} else {
|
|
store.clear().map_err(internal_server_error)?;
|
|
}
|
|
Ok(Response::builder(Status::NO_CONTENT).build())
|
|
}
|
|
(path, "POST") if path.starts_with("/store") => {
|
|
if read_only {
|
|
return Err(the_server_is_read_only());
|
|
}
|
|
let content_type =
|
|
content_type(request).ok_or_else(|| bad_request("No Content-Type given"))?;
|
|
if let Some(target) = store_target(request)? {
|
|
let format = RdfFormat::from_media_type(&content_type)
|
|
.ok_or_else(|| unsupported_media_type(&content_type))?;
|
|
let new = assert_that_graph_exists(&store, &target).is_ok();
|
|
web_load_graph(&store, request, format, &GraphName::from(target))?;
|
|
Ok(Response::builder(if new {
|
|
Status::CREATED
|
|
} else {
|
|
Status::NO_CONTENT
|
|
})
|
|
.build())
|
|
} else {
|
|
let format = RdfFormat::from_media_type(&content_type)
|
|
.ok_or_else(|| unsupported_media_type(&content_type))?;
|
|
if format.supports_datasets() {
|
|
web_load_dataset(&store, request, format)?;
|
|
Ok(Response::builder(Status::NO_CONTENT).build())
|
|
} else {
|
|
let graph =
|
|
resolve_with_base(request, &format!("/store/{:x}", random::<u128>()))?;
|
|
web_load_graph(&store, request, format, &graph.clone().into())?;
|
|
Ok(Response::builder(Status::CREATED)
|
|
.with_header(HeaderName::LOCATION, graph.into_string())
|
|
.unwrap()
|
|
.build())
|
|
}
|
|
}
|
|
}
|
|
(path, "HEAD") if path.starts_with("/store") => {
|
|
if let Some(target) = store_target(request)? {
|
|
assert_that_graph_exists(&store, &target)?;
|
|
}
|
|
Ok(Response::builder(Status::OK).build())
|
|
}
|
|
_ => Err((
|
|
Status::NOT_FOUND,
|
|
format!(
|
|
"{} {} is not supported by this server",
|
|
request.method(),
|
|
request.url().path()
|
|
),
|
|
)),
|
|
}
|
|
}
|
|
|
|
fn base_url(request: &Request) -> String {
|
|
let mut url = request.url().clone();
|
|
url.set_query(None);
|
|
url.set_fragment(None);
|
|
url.into()
|
|
}
|
|
|
|
fn resolve_with_base(request: &Request, url: &str) -> Result<NamedNode, HttpError> {
|
|
Ok(Iri::parse(base_url(request))
|
|
.map_err(bad_request)?
|
|
.resolve(url)
|
|
.map_err(bad_request)?
|
|
.into())
|
|
}
|
|
|
|
fn url_query(request: &Request) -> &[u8] {
|
|
request.url().query().unwrap_or("").as_bytes()
|
|
}
|
|
|
|
fn url_query_parameter<'a>(request: &'a Request, param: &str) -> Option<Cow<'a, str>> {
|
|
request
|
|
.url()
|
|
.query_pairs()
|
|
.find(|(k, _)| k == param)
|
|
.map(|(_, v)| v)
|
|
}
|
|
|
|
fn configure_and_evaluate_sparql_query(
|
|
store: &Store,
|
|
encoded: &[&[u8]],
|
|
mut query: Option<String>,
|
|
request: &Request,
|
|
) -> Result<Response, HttpError> {
|
|
let mut default_graph_uris = Vec::new();
|
|
let mut named_graph_uris = Vec::new();
|
|
let mut use_default_graph_as_union = false;
|
|
for encoded in encoded {
|
|
for (k, v) in form_urlencoded::parse(encoded) {
|
|
match k.as_ref() {
|
|
"query" => {
|
|
if query.is_some() {
|
|
return Err(bad_request("Multiple query parameters provided"));
|
|
}
|
|
query = Some(v.into_owned())
|
|
}
|
|
"default-graph-uri" => default_graph_uris.push(v.into_owned()),
|
|
"union-default-graph" => use_default_graph_as_union = true,
|
|
"named-graph-uri" => named_graph_uris.push(v.into_owned()),
|
|
_ => (),
|
|
}
|
|
}
|
|
}
|
|
let query = query.ok_or_else(|| bad_request("You should set the 'query' parameter"))?;
|
|
evaluate_sparql_query(
|
|
store,
|
|
&query,
|
|
use_default_graph_as_union,
|
|
default_graph_uris,
|
|
named_graph_uris,
|
|
request,
|
|
)
|
|
}
|
|
|
|
fn evaluate_sparql_query(
|
|
store: &Store,
|
|
query: &str,
|
|
use_default_graph_as_union: bool,
|
|
default_graph_uris: Vec<String>,
|
|
named_graph_uris: Vec<String>,
|
|
request: &Request,
|
|
) -> Result<Response, HttpError> {
|
|
let mut query = Query::parse(query, Some(&base_url(request))).map_err(bad_request)?;
|
|
|
|
if use_default_graph_as_union {
|
|
if !default_graph_uris.is_empty() || !named_graph_uris.is_empty() {
|
|
return Err(bad_request(
|
|
"default-graph-uri or named-graph-uri and union-default-graph should not be set at the same time"
|
|
));
|
|
}
|
|
query.dataset_mut().set_default_graph_as_union()
|
|
} else if !default_graph_uris.is_empty() || !named_graph_uris.is_empty() {
|
|
query.dataset_mut().set_default_graph(
|
|
default_graph_uris
|
|
.into_iter()
|
|
.map(|e| Ok(NamedNode::new(e)?.into()))
|
|
.collect::<Result<Vec<GraphName>, IriParseError>>()
|
|
.map_err(bad_request)?,
|
|
);
|
|
query.dataset_mut().set_available_named_graphs(
|
|
named_graph_uris
|
|
.into_iter()
|
|
.map(|e| Ok(NamedNode::new(e)?.into()))
|
|
.collect::<Result<Vec<NamedOrBlankNode>, IriParseError>>()
|
|
.map_err(bad_request)?,
|
|
);
|
|
}
|
|
|
|
let results = store.query(query).map_err(internal_server_error)?;
|
|
match results {
|
|
QueryResults::Solutions(solutions) => {
|
|
let format = query_results_content_negotiation(request)?;
|
|
ReadForWrite::build_response(
|
|
move |w| {
|
|
Ok((
|
|
QueryResultsSerializer::from_format(format)
|
|
.serialize_solutions_to_write(w, solutions.variables().to_vec())?,
|
|
solutions,
|
|
))
|
|
},
|
|
|(mut writer, mut solutions)| {
|
|
Ok(if let Some(solution) = solutions.next() {
|
|
writer.write(&solution?)?;
|
|
Some((writer, solutions))
|
|
} else {
|
|
writer.finish()?;
|
|
None
|
|
})
|
|
},
|
|
format.media_type(),
|
|
)
|
|
}
|
|
QueryResults::Boolean(result) => {
|
|
let format = query_results_content_negotiation(request)?;
|
|
let mut body = Vec::new();
|
|
QueryResultsSerializer::from_format(format)
|
|
.serialize_boolean_to_write(&mut body, result)
|
|
.map_err(internal_server_error)?;
|
|
Ok(Response::builder(Status::OK)
|
|
.with_header(HeaderName::CONTENT_TYPE, format.media_type())
|
|
.unwrap()
|
|
.with_body(body))
|
|
}
|
|
QueryResults::Graph(triples) => {
|
|
let format = rdf_content_negotiation(request)?;
|
|
ReadForWrite::build_response(
|
|
move |w| {
|
|
Ok((
|
|
RdfSerializer::from_format(format).serialize_to_write(w),
|
|
triples,
|
|
))
|
|
},
|
|
|(mut writer, mut triples)| {
|
|
Ok(if let Some(t) = triples.next() {
|
|
writer.write_triple(&t?)?;
|
|
Some((writer, triples))
|
|
} else {
|
|
writer.finish()?;
|
|
None
|
|
})
|
|
},
|
|
format.media_type(),
|
|
)
|
|
}
|
|
}
|
|
}
|
|
|
|
fn configure_and_evaluate_sparql_update(
|
|
store: &Store,
|
|
encoded: &[&[u8]],
|
|
mut update: Option<String>,
|
|
request: &Request,
|
|
) -> Result<Response, HttpError> {
|
|
let mut use_default_graph_as_union = false;
|
|
let mut default_graph_uris = Vec::new();
|
|
let mut named_graph_uris = Vec::new();
|
|
for encoded in encoded {
|
|
for (k, v) in form_urlencoded::parse(encoded) {
|
|
match k.as_ref() {
|
|
"update" => {
|
|
if update.is_some() {
|
|
return Err(bad_request("Multiple update parameters provided"));
|
|
}
|
|
update = Some(v.into_owned())
|
|
}
|
|
"using-graph-uri" => default_graph_uris.push(v.into_owned()),
|
|
"using-union-graph" => use_default_graph_as_union = true,
|
|
"using-named-graph-uri" => named_graph_uris.push(v.into_owned()),
|
|
_ => (),
|
|
}
|
|
}
|
|
}
|
|
let update = update.ok_or_else(|| bad_request("You should set the 'update' parameter"))?;
|
|
evaluate_sparql_update(
|
|
store,
|
|
&update,
|
|
use_default_graph_as_union,
|
|
default_graph_uris,
|
|
named_graph_uris,
|
|
request,
|
|
)
|
|
}
|
|
|
|
fn evaluate_sparql_update(
|
|
store: &Store,
|
|
update: &str,
|
|
use_default_graph_as_union: bool,
|
|
default_graph_uris: Vec<String>,
|
|
named_graph_uris: Vec<String>,
|
|
request: &Request,
|
|
) -> Result<Response, HttpError> {
|
|
let mut update =
|
|
Update::parse(update, Some(base_url(request).as_str())).map_err(bad_request)?;
|
|
|
|
if use_default_graph_as_union {
|
|
if !default_graph_uris.is_empty() || !named_graph_uris.is_empty() {
|
|
return Err(bad_request(
|
|
"using-graph-uri or using-named-graph-uri and using-union-graph should not be set at the same time"
|
|
));
|
|
}
|
|
for using in update.using_datasets_mut() {
|
|
if !using.is_default_dataset() {
|
|
return Err(bad_request(
|
|
"using-union-graph must not be used with a SPARQL UPDATE containing USING",
|
|
));
|
|
}
|
|
using.set_default_graph_as_union();
|
|
}
|
|
} else if !default_graph_uris.is_empty() || !named_graph_uris.is_empty() {
|
|
let default_graph_uris = default_graph_uris
|
|
.into_iter()
|
|
.map(|e| Ok(NamedNode::new(e)?.into()))
|
|
.collect::<Result<Vec<GraphName>, IriParseError>>()
|
|
.map_err(bad_request)?;
|
|
let named_graph_uris = named_graph_uris
|
|
.into_iter()
|
|
.map(|e| Ok(NamedNode::new(e)?.into()))
|
|
.collect::<Result<Vec<NamedOrBlankNode>, IriParseError>>()
|
|
.map_err(bad_request)?;
|
|
for using in update.using_datasets_mut() {
|
|
if !using.is_default_dataset() {
|
|
return Err(bad_request(
|
|
"using-graph-uri and using-named-graph-uri must not be used with a SPARQL UPDATE containing USING",
|
|
));
|
|
}
|
|
using.set_default_graph(default_graph_uris.clone());
|
|
using.set_available_named_graphs(named_graph_uris.clone());
|
|
}
|
|
}
|
|
store.update(update).map_err(internal_server_error)?;
|
|
Ok(Response::builder(Status::NO_CONTENT).build())
|
|
}
|
|
|
|
fn store_target(request: &Request) -> Result<Option<NamedGraphName>, HttpError> {
|
|
if request.url().path() == "/store" {
|
|
let mut graph = None;
|
|
let mut default = false;
|
|
for (k, v) in request.url().query_pairs() {
|
|
match k.as_ref() {
|
|
"graph" => graph = Some(v.into_owned()),
|
|
"default" => default = true,
|
|
_ => continue,
|
|
}
|
|
}
|
|
if let Some(graph) = graph {
|
|
if default {
|
|
Err(bad_request(
|
|
"Both graph and default parameters should not be set at the same time",
|
|
))
|
|
} else {
|
|
Ok(Some(NamedGraphName::NamedNode(resolve_with_base(
|
|
request, &graph,
|
|
)?)))
|
|
}
|
|
} else if default {
|
|
Ok(Some(NamedGraphName::DefaultGraph))
|
|
} else {
|
|
Ok(None)
|
|
}
|
|
} else {
|
|
Ok(Some(NamedGraphName::NamedNode(resolve_with_base(
|
|
request, "",
|
|
)?)))
|
|
}
|
|
}
|
|
|
|
fn assert_that_graph_exists(store: &Store, target: &NamedGraphName) -> Result<(), HttpError> {
|
|
if match target {
|
|
NamedGraphName::DefaultGraph => true,
|
|
NamedGraphName::NamedNode(target) => store
|
|
.contains_named_graph(target)
|
|
.map_err(internal_server_error)?,
|
|
} {
|
|
Ok(())
|
|
} else {
|
|
Err((
|
|
Status::NOT_FOUND,
|
|
format!(
|
|
"The graph {} does not exists",
|
|
GraphName::from(target.clone())
|
|
),
|
|
))
|
|
}
|
|
}
|
|
|
|
#[derive(Eq, PartialEq, Debug, Clone, Hash)]
|
|
enum NamedGraphName {
|
|
NamedNode(NamedNode),
|
|
DefaultGraph,
|
|
}
|
|
|
|
impl From<NamedGraphName> for GraphName {
|
|
fn from(graph_name: NamedGraphName) -> Self {
|
|
match graph_name {
|
|
NamedGraphName::NamedNode(node) => node.into(),
|
|
NamedGraphName::DefaultGraph => Self::DefaultGraph,
|
|
}
|
|
}
|
|
}
|
|
|
|
fn rdf_content_negotiation(request: &Request) -> Result<RdfFormat, HttpError> {
|
|
content_negotiation(
|
|
request,
|
|
RdfFormat::from_media_type,
|
|
RdfFormat::NQuads,
|
|
&[
|
|
("application", RdfFormat::NQuads),
|
|
("text", RdfFormat::NQuads),
|
|
],
|
|
"application/n-quads or text/turtle",
|
|
)
|
|
}
|
|
|
|
fn query_results_content_negotiation(request: &Request) -> Result<QueryResultsFormat, HttpError> {
|
|
content_negotiation(
|
|
request,
|
|
QueryResultsFormat::from_media_type,
|
|
QueryResultsFormat::Json,
|
|
&[
|
|
("application", QueryResultsFormat::Json),
|
|
("text", QueryResultsFormat::Json),
|
|
],
|
|
"application/sparql-results+json or text/tsv",
|
|
)
|
|
}
|
|
|
|
fn content_negotiation<F: Copy>(
|
|
request: &Request,
|
|
parse: impl Fn(&str) -> Option<F>,
|
|
default: F,
|
|
default_by_base: &[(&str, F)],
|
|
example: &str,
|
|
) -> Result<F, HttpError> {
|
|
let default_value = HeaderValue::default();
|
|
let header = request
|
|
.header(&HeaderName::ACCEPT)
|
|
.unwrap_or(&default_value)
|
|
.to_str()
|
|
.map_err(|_| bad_request("The Accept header should be a valid ASCII string"))?;
|
|
|
|
if header.is_empty() {
|
|
return Ok(default);
|
|
}
|
|
let mut result = None;
|
|
let mut result_score = 0_f32;
|
|
for mut possible in header.split(',') {
|
|
let mut score = 1.;
|
|
if let Some((possible_type, last_parameter)) = possible.rsplit_once(';') {
|
|
if let Some((name, value)) = last_parameter.split_once('=') {
|
|
if name.trim().eq_ignore_ascii_case("q") {
|
|
score = f32::from_str(value.trim()).map_err(|_| {
|
|
bad_request(format!("Invalid Accept media type score: {value}"))
|
|
})?;
|
|
possible = possible_type;
|
|
}
|
|
}
|
|
}
|
|
if score <= result_score {
|
|
continue;
|
|
}
|
|
let (possible_base, possible_sub) = possible
|
|
.split_once(';')
|
|
.unwrap_or((possible, ""))
|
|
.0
|
|
.split_once('/')
|
|
.ok_or_else(|| bad_request(format!("Invalid media type: '{possible}'")))?;
|
|
let possible_base = possible_base.trim();
|
|
let possible_sub = possible_sub.trim();
|
|
|
|
let mut format = None;
|
|
if possible_base == "*" && possible_sub == "*" {
|
|
format = Some(default);
|
|
} else if possible_sub == "*" {
|
|
for (base, sub_format) in default_by_base {
|
|
if *base == possible_base {
|
|
format = Some(*sub_format);
|
|
}
|
|
}
|
|
} else {
|
|
format = parse(possible);
|
|
}
|
|
if let Some(format) = format {
|
|
result = Some(format);
|
|
result_score = score;
|
|
}
|
|
}
|
|
|
|
result.ok_or_else(|| {
|
|
(
|
|
Status::NOT_ACCEPTABLE,
|
|
format!("The accept header does not provide any accepted format like {example}"),
|
|
)
|
|
})
|
|
}
|
|
|
|
fn content_type(request: &Request) -> Option<String> {
|
|
let value = request.header(&HeaderName::CONTENT_TYPE)?.to_str().ok()?;
|
|
Some(
|
|
value
|
|
.split_once(';')
|
|
.map_or(value, |(b, _)| b)
|
|
.trim()
|
|
.to_ascii_lowercase(),
|
|
)
|
|
}
|
|
|
|
fn web_load_graph(
|
|
store: &Store,
|
|
request: &mut Request,
|
|
format: RdfFormat,
|
|
to_graph_name: &GraphName,
|
|
) -> Result<(), HttpError> {
|
|
let base_iri = if let GraphName::NamedNode(graph_name) = to_graph_name {
|
|
Some(graph_name.as_str())
|
|
} else {
|
|
None
|
|
};
|
|
let mut parser = RdfParser::from_format(format)
|
|
.without_named_graphs()
|
|
.with_default_graph(to_graph_name.clone());
|
|
if let Some(base_iri) = base_iri {
|
|
parser = parser.with_base_iri(base_iri).map_err(bad_request)?;
|
|
}
|
|
if url_query_parameter(request, "no_transaction").is_some() {
|
|
web_bulk_loader(store, request).load_from_read(parser, request.body_mut())
|
|
} else {
|
|
store.load_from_read(parser, request.body_mut())
|
|
}
|
|
.map_err(loader_to_http_error)
|
|
}
|
|
|
|
fn web_load_dataset(
|
|
store: &Store,
|
|
request: &mut Request,
|
|
format: RdfFormat,
|
|
) -> Result<(), HttpError> {
|
|
if url_query_parameter(request, "no_transaction").is_some() {
|
|
web_bulk_loader(store, request).load_from_read(format, request.body_mut())
|
|
} else {
|
|
store.load_from_read(format, request.body_mut())
|
|
}
|
|
.map_err(loader_to_http_error)
|
|
}
|
|
|
|
fn web_bulk_loader(store: &Store, request: &Request) -> BulkLoader {
|
|
let start = Instant::now();
|
|
let mut loader = store.bulk_loader().on_progress(move |size| {
|
|
let elapsed = start.elapsed();
|
|
eprintln!(
|
|
"{} triples loaded in {}s ({} t/s)",
|
|
size,
|
|
elapsed.as_secs(),
|
|
((size as f64) / elapsed.as_secs_f64()).round()
|
|
)
|
|
});
|
|
if url_query_parameter(request, "lenient").is_some() {
|
|
loader = loader.on_parse_error(move |e| {
|
|
eprintln!("Parsing error: {e}");
|
|
Ok(())
|
|
})
|
|
}
|
|
loader
|
|
}
|
|
|
|
fn error(status: Status, message: impl fmt::Display) -> Response {
|
|
Response::builder(status)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/plain; charset=utf-8")
|
|
.unwrap()
|
|
.with_body(message.to_string())
|
|
}
|
|
|
|
fn bad_request(message: impl fmt::Display) -> HttpError {
|
|
(Status::BAD_REQUEST, message.to_string())
|
|
}
|
|
|
|
fn the_server_is_read_only() -> HttpError {
|
|
(Status::FORBIDDEN, "The server is read-only".into())
|
|
}
|
|
|
|
fn unsupported_media_type(content_type: &str) -> HttpError {
|
|
(
|
|
Status::UNSUPPORTED_MEDIA_TYPE,
|
|
format!("No supported content Content-Type given: {content_type}"),
|
|
)
|
|
}
|
|
|
|
fn internal_server_error(message: impl fmt::Display) -> HttpError {
|
|
eprintln!("Internal server error: {message}");
|
|
(Status::INTERNAL_SERVER_ERROR, message.to_string())
|
|
}
|
|
|
|
fn loader_to_http_error(e: LoaderError) -> HttpError {
|
|
match e {
|
|
LoaderError::Parsing(e) => bad_request(e),
|
|
LoaderError::Storage(e) => internal_server_error(e),
|
|
LoaderError::InvalidBaseIri { .. } => bad_request(e),
|
|
}
|
|
}
|
|
|
|
/// Hacky tool to allow implementing read on top of a write loop
|
|
struct ReadForWrite<O, U: (Fn(O) -> io::Result<Option<O>>)> {
|
|
buffer: Rc<RefCell<Vec<u8>>>,
|
|
position: usize,
|
|
add_more_data: U,
|
|
state: Option<O>,
|
|
}
|
|
|
|
impl<O: 'static, U: (Fn(O) -> io::Result<Option<O>>) + 'static> ReadForWrite<O, U> {
|
|
fn build_response(
|
|
initial_state_builder: impl FnOnce(ReadForWriteWriter) -> io::Result<O>,
|
|
add_more_data: U,
|
|
content_type: &'static str,
|
|
) -> Result<Response, HttpError> {
|
|
let buffer = Rc::new(RefCell::new(Vec::new()));
|
|
let state = initial_state_builder(ReadForWriteWriter {
|
|
buffer: Rc::clone(&buffer),
|
|
})
|
|
.map_err(internal_server_error)?;
|
|
Ok(Response::builder(Status::OK)
|
|
.with_header(HeaderName::CONTENT_TYPE, content_type)
|
|
.map_err(internal_server_error)?
|
|
.with_body(Body::from_read(Self {
|
|
buffer,
|
|
position: 0,
|
|
add_more_data,
|
|
state: Some(state),
|
|
})))
|
|
}
|
|
}
|
|
|
|
impl<O, U: (Fn(O) -> io::Result<Option<O>>)> Read for ReadForWrite<O, U> {
|
|
fn read(&mut self, buf: &mut [u8]) -> io::Result<usize> {
|
|
while self.position == self.buffer.borrow().len() {
|
|
// We read more data
|
|
if let Some(state) = self.state.take() {
|
|
self.buffer.borrow_mut().clear();
|
|
self.position = 0;
|
|
self.state = match (self.add_more_data)(state) {
|
|
Ok(state) => state,
|
|
Err(e) => {
|
|
eprintln!("Internal server error while streaming results: {e}");
|
|
self.buffer
|
|
.borrow_mut()
|
|
.write_all(e.to_string().as_bytes())?;
|
|
None
|
|
}
|
|
}
|
|
} else {
|
|
return Ok(0); // End
|
|
}
|
|
}
|
|
let buffer = self.buffer.borrow();
|
|
let len = min(buffer.len() - self.position, buf.len());
|
|
buf[..len].copy_from_slice(&buffer[self.position..self.position + len]);
|
|
self.position += len;
|
|
Ok(len)
|
|
}
|
|
}
|
|
|
|
struct ReadForWriteWriter {
|
|
buffer: Rc<RefCell<Vec<u8>>>,
|
|
}
|
|
|
|
impl Write for ReadForWriteWriter {
|
|
fn write(&mut self, buf: &[u8]) -> io::Result<usize> {
|
|
self.buffer.borrow_mut().write(buf)
|
|
}
|
|
|
|
fn flush(&mut self) -> io::Result<()> {
|
|
Ok(())
|
|
}
|
|
|
|
fn write_all(&mut self, buf: &[u8]) -> io::Result<()> {
|
|
self.buffer.borrow_mut().write_all(buf)
|
|
}
|
|
}
|
|
|
|
fn close_file_writer(writer: BufWriter<File>) -> io::Result<()> {
|
|
let mut file = writer
|
|
.into_inner()
|
|
.map_err(io::IntoInnerError::into_error)?;
|
|
file.flush()?;
|
|
file.sync_all()
|
|
}
|
|
|
|
#[cfg(target_os = "linux")]
|
|
fn systemd_notify_ready() -> io::Result<()> {
|
|
if let Some(path) = env::var_os("NOTIFY_SOCKET") {
|
|
UnixDatagram::unbound()?.send_to(b"READY=1", path)?;
|
|
}
|
|
Ok(())
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod tests {
|
|
#![allow(clippy::panic_in_result_fn)]
|
|
|
|
use super::*;
|
|
use anyhow::Result;
|
|
use assert_cmd::Command;
|
|
use assert_fs::prelude::*;
|
|
use assert_fs::{NamedTempFile, TempDir};
|
|
use flate2::write::GzEncoder;
|
|
use flate2::Compression;
|
|
use oxhttp::model::Method;
|
|
use predicates::prelude::*;
|
|
use std::fs::remove_dir_all;
|
|
use std::io::read_to_string;
|
|
|
|
fn cli_command() -> Result<Command> {
|
|
Ok(Command::from_std(
|
|
escargot::CargoBuild::new()
|
|
.bin("oxigraph")
|
|
.manifest_path(format!("{}/Cargo.toml", env!("CARGO_MANIFEST_DIR")))
|
|
.run()?
|
|
.command(),
|
|
))
|
|
}
|
|
|
|
fn initialized_cli_store(data: &'static str) -> Result<TempDir> {
|
|
let store_dir = TempDir::new()?;
|
|
cli_command()?
|
|
.arg("load")
|
|
.arg("--location")
|
|
.arg(store_dir.path())
|
|
.arg("--format")
|
|
.arg("trig")
|
|
.write_stdin(data)
|
|
.assert()
|
|
.success();
|
|
Ok(store_dir)
|
|
}
|
|
|
|
fn assert_cli_state(store_dir: &TempDir, data: &'static str) -> Result<()> {
|
|
cli_command()?
|
|
.arg("dump")
|
|
.arg("--location")
|
|
.arg(store_dir.path())
|
|
.arg("--format")
|
|
.arg("nq")
|
|
.assert()
|
|
.stdout(data)
|
|
.success();
|
|
Ok(())
|
|
}
|
|
|
|
#[test]
|
|
fn cli_help() -> Result<()> {
|
|
cli_command()?
|
|
.assert()
|
|
.failure()
|
|
.stdout("")
|
|
.stderr(predicate::str::starts_with("Oxigraph"));
|
|
Ok(())
|
|
}
|
|
|
|
#[test]
|
|
fn cli_load_optimize_and_dump_graph() -> Result<()> {
|
|
let store_dir = TempDir::new()?;
|
|
let input_file = NamedTempFile::new("input.ttl")?;
|
|
input_file.write_str("<s> <http://example.com/p> <http://example.com/o> .")?;
|
|
cli_command()?
|
|
.arg("load")
|
|
.arg("--location")
|
|
.arg(store_dir.path())
|
|
.arg("--file")
|
|
.arg(input_file.path())
|
|
.arg("--base")
|
|
.arg("http://example.com/")
|
|
.assert()
|
|
.success();
|
|
|
|
cli_command()?
|
|
.arg("optimize")
|
|
.arg("--location")
|
|
.arg(store_dir.path())
|
|
.assert()
|
|
.success();
|
|
|
|
let output_file = NamedTempFile::new("output.nt")?;
|
|
cli_command()?
|
|
.arg("dump")
|
|
.arg("--location")
|
|
.arg(store_dir.path())
|
|
.arg("--file")
|
|
.arg(output_file.path())
|
|
.arg("--graph")
|
|
.arg("default")
|
|
.assert()
|
|
.success();
|
|
output_file
|
|
.assert("<http://example.com/s> <http://example.com/p> <http://example.com/o> .\n");
|
|
Ok(())
|
|
}
|
|
|
|
#[test]
|
|
fn cli_load_and_dump_dataset() -> Result<()> {
|
|
let store_dir = TempDir::new()?;
|
|
let input_file = NamedTempFile::new("input.nq")?;
|
|
input_file
|
|
.write_str("<http://example.com/s> <http://example.com/p> <http://example.com/o> <http://example.com/g> .")?;
|
|
cli_command()?
|
|
.arg("load")
|
|
.arg("--location")
|
|
.arg(store_dir.path())
|
|
.arg("--file")
|
|
.arg(input_file.path())
|
|
.assert()
|
|
.success();
|
|
|
|
let output_file = NamedTempFile::new("output.nq")?;
|
|
cli_command()?
|
|
.arg("dump")
|
|
.arg("--location")
|
|
.arg(store_dir.path())
|
|
.arg("--file")
|
|
.arg(output_file.path())
|
|
.assert()
|
|
.success();
|
|
output_file
|
|
.assert("<http://example.com/s> <http://example.com/p> <http://example.com/o> <http://example.com/g> .\n");
|
|
Ok(())
|
|
}
|
|
|
|
#[test]
|
|
fn cli_load_gzip_dataset() -> Result<()> {
|
|
let store_dir = TempDir::new()?;
|
|
let file = NamedTempFile::new("sample.nq.gz")?;
|
|
let mut encoder = GzEncoder::new(Vec::new(), Compression::default());
|
|
encoder
|
|
.write_all(b"<http://example.com/s> <http://example.com/p> <http://example.com/o> .")?;
|
|
file.write_binary(&encoder.finish()?)?;
|
|
cli_command()?
|
|
.arg("load")
|
|
.arg("-l")
|
|
.arg(store_dir.path())
|
|
.arg("-f")
|
|
.arg(file.path())
|
|
.assert()
|
|
.success();
|
|
|
|
cli_command()?
|
|
.arg("dump")
|
|
.arg("-l")
|
|
.arg(store_dir.path())
|
|
.arg("--format")
|
|
.arg("nq")
|
|
.assert()
|
|
.success()
|
|
.stdout("<http://example.com/s> <http://example.com/p> <http://example.com/o> .\n");
|
|
Ok(())
|
|
}
|
|
|
|
#[test]
|
|
fn cli_load_and_dump_named_graph() -> Result<()> {
|
|
let store_dir = TempDir::new()?;
|
|
let input_file = NamedTempFile::new("input.nt")?;
|
|
input_file.write_str(
|
|
"<http://example.com/s> <http://example.com/p> <http://example.com/o> .\n",
|
|
)?;
|
|
cli_command()?
|
|
.arg("load")
|
|
.arg("-l")
|
|
.arg(store_dir.path())
|
|
.arg("-f")
|
|
.arg(input_file.path())
|
|
.arg("--graph")
|
|
.arg("http://example.com/g")
|
|
.assert()
|
|
.success();
|
|
|
|
let output_file = NamedTempFile::new("output.nt")?;
|
|
cli_command()?
|
|
.arg("dump")
|
|
.arg("-l")
|
|
.arg(store_dir.path())
|
|
.arg("-f")
|
|
.arg(output_file.path())
|
|
.arg("--graph")
|
|
.arg("http://example.com/g")
|
|
.assert()
|
|
.success();
|
|
output_file
|
|
.assert("<http://example.com/s> <http://example.com/p> <http://example.com/o> .\n");
|
|
Ok(())
|
|
}
|
|
|
|
#[test]
|
|
fn cli_load_and_dump_with_format() -> Result<()> {
|
|
let store_dir = TempDir::new()?;
|
|
let input_file = NamedTempFile::new("input")?;
|
|
input_file
|
|
.write_str("<http://example.com/s> <http://example.com/p> <http://example.com/o> .")?;
|
|
cli_command()?
|
|
.arg("load")
|
|
.arg("-l")
|
|
.arg(store_dir.path())
|
|
.arg("-f")
|
|
.arg(input_file.path())
|
|
.arg("--format")
|
|
.arg("nt")
|
|
.assert()
|
|
.success();
|
|
|
|
let output_file = NamedTempFile::new("output")?;
|
|
cli_command()?
|
|
.arg("dump")
|
|
.arg("--location")
|
|
.arg(store_dir.path())
|
|
.arg("--file")
|
|
.arg(output_file.path())
|
|
.arg("--graph")
|
|
.arg("default")
|
|
.arg("--format")
|
|
.arg("nt")
|
|
.assert()
|
|
.success();
|
|
output_file
|
|
.assert("<http://example.com/s> <http://example.com/p> <http://example.com/o> .\n");
|
|
Ok(())
|
|
}
|
|
|
|
#[test]
|
|
fn cli_load_from_stdin_and_dump_to_stdout() -> Result<()> {
|
|
let store_dir = TempDir::new()?;
|
|
cli_command()?
|
|
.arg("load")
|
|
.arg("--location")
|
|
.arg(store_dir.path())
|
|
.arg("--format")
|
|
.arg("nq")
|
|
.write_stdin("<http://example.com/s> <http://example.com/p> <http://example.com/o> .")
|
|
.assert()
|
|
.success();
|
|
|
|
cli_command()?
|
|
.arg("dump")
|
|
.arg("--location")
|
|
.arg(store_dir.path())
|
|
.arg("--format")
|
|
.arg("nq")
|
|
.assert()
|
|
.success()
|
|
.stdout("<http://example.com/s> <http://example.com/p> <http://example.com/o> .\n");
|
|
Ok(())
|
|
}
|
|
|
|
#[test]
|
|
fn cli_backup() -> Result<()> {
|
|
let store_dir = initialized_cli_store(
|
|
"<http://example.com/s> <http://example.com/p> <http://example.com/o> .",
|
|
)?;
|
|
|
|
let backup_dir = TempDir::new()?;
|
|
remove_dir_all(backup_dir.path())?; // The directory should not exist yet
|
|
cli_command()?
|
|
.arg("backup")
|
|
.arg("--location")
|
|
.arg(store_dir.path())
|
|
.arg("--destination")
|
|
.arg(backup_dir.path())
|
|
.assert()
|
|
.success();
|
|
|
|
assert_cli_state(
|
|
&store_dir,
|
|
"<http://example.com/s> <http://example.com/p> <http://example.com/o> .\n",
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn cli_ask_query_inline() -> Result<()> {
|
|
let store_dir = initialized_cli_store(
|
|
"<http://example.com/s> <http://example.com/p> <http://example.com/o> .",
|
|
)?;
|
|
cli_command()?
|
|
.arg("query")
|
|
.arg("--location")
|
|
.arg(store_dir.path())
|
|
.arg("--query")
|
|
.arg("ASK { <s> <p> <o> }")
|
|
.arg("--query-base")
|
|
.arg("http://example.com/")
|
|
.arg("--results-format")
|
|
.arg("csv")
|
|
.assert()
|
|
.stdout("true")
|
|
.success();
|
|
Ok(())
|
|
}
|
|
|
|
#[test]
|
|
fn cli_construct_query_stdin() -> Result<()> {
|
|
let store_dir = initialized_cli_store(
|
|
"<http://example.com/s> <http://example.com/p> <http://example.com/o> .",
|
|
)?;
|
|
cli_command()?
|
|
.arg("query")
|
|
.arg("--location")
|
|
.arg(store_dir.path())
|
|
.arg("--query-base")
|
|
.arg("http://example.com/")
|
|
.arg("--results-format")
|
|
.arg("nt")
|
|
.write_stdin("CONSTRUCT { ?s ?p ?o } WHERE { ?s ?p ?o }")
|
|
.assert()
|
|
.stdout("<http://example.com/s> <http://example.com/p> <http://example.com/o> .\n")
|
|
.success();
|
|
Ok(())
|
|
}
|
|
|
|
#[test]
|
|
fn cli_select_query_file() -> Result<()> {
|
|
let store_dir = initialized_cli_store(
|
|
"<http://example.com/s> <http://example.com/p> <http://example.com/o> .",
|
|
)?;
|
|
let input_file = NamedTempFile::new("input.rq")?;
|
|
input_file.write_str("SELECT ?s WHERE { ?s ?p ?o }")?;
|
|
let output_file = NamedTempFile::new("output.tsv")?;
|
|
cli_command()?
|
|
.arg("query")
|
|
.arg("--location")
|
|
.arg(store_dir.path())
|
|
.arg("--query-file")
|
|
.arg(input_file.path())
|
|
.arg("--results-file")
|
|
.arg(output_file.path())
|
|
.assert()
|
|
.success();
|
|
output_file.assert("?s\n<http://example.com/s>\n");
|
|
Ok(())
|
|
}
|
|
|
|
#[test]
|
|
fn cli_ask_update_inline() -> Result<()> {
|
|
let store_dir = TempDir::new()?;
|
|
cli_command()?
|
|
.arg("update")
|
|
.arg("--location")
|
|
.arg(store_dir.path())
|
|
.arg("--update")
|
|
.arg("INSERT DATA { <s> <p> <o> }")
|
|
.arg("--update-base")
|
|
.arg("http://example.com/")
|
|
.assert()
|
|
.success();
|
|
assert_cli_state(
|
|
&store_dir,
|
|
"<http://example.com/s> <http://example.com/p> <http://example.com/o> .\n",
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn cli_construct_update_stdin() -> Result<()> {
|
|
let store_dir = TempDir::new()?;
|
|
cli_command()?
|
|
.arg("update")
|
|
.arg("--location")
|
|
.arg(store_dir.path())
|
|
.arg("--update-base")
|
|
.arg("http://example.com/")
|
|
.write_stdin("INSERT DATA { <s> <p> <o> }")
|
|
.assert()
|
|
.success();
|
|
assert_cli_state(
|
|
&store_dir,
|
|
"<http://example.com/s> <http://example.com/p> <http://example.com/o> .\n",
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn cli_update_file() -> Result<()> {
|
|
let store_dir = TempDir::new()?;
|
|
let input_file = NamedTempFile::new("input.rq")?;
|
|
input_file.write_str(
|
|
"INSERT DATA { <http://example.com/s> <http://example.com/p> <http://example.com/o> }",
|
|
)?;
|
|
cli_command()?
|
|
.arg("update")
|
|
.arg("--location")
|
|
.arg(store_dir.path())
|
|
.arg("--update-file")
|
|
.arg(input_file.path())
|
|
.assert()
|
|
.success();
|
|
assert_cli_state(
|
|
&store_dir,
|
|
"<http://example.com/s> <http://example.com/p> <http://example.com/o> .\n",
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn cli_convert_file() -> Result<()> {
|
|
let input_file = NamedTempFile::new("input.ttl")?;
|
|
input_file.write_str("@prefix schema: <http://schema.org/> .\n<http://example.com#me> a schema:Person ;\n\tschema:name \"Foo Bar\"@en .\n")?;
|
|
let output_file = NamedTempFile::new("output.rdf")?;
|
|
cli_command()?
|
|
.arg("convert")
|
|
.arg("--from-file")
|
|
.arg(input_file.path())
|
|
.arg("--from-base")
|
|
.arg("http://example.com/")
|
|
.arg("--to-file")
|
|
.arg(output_file.path())
|
|
.assert()
|
|
.success();
|
|
output_file
|
|
.assert("<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n<rdf:RDF xmlns:schema=\"http://schema.org/\" xmlns:rdf=\"http://www.w3.org/1999/02/22-rdf-syntax-ns#\">\n\t<schema:Person rdf:about=\"http://example.com#me\">\n\t\t<schema:name xml:lang=\"en\">Foo Bar</schema:name>\n\t</schema:Person>\n</rdf:RDF>");
|
|
Ok(())
|
|
}
|
|
|
|
#[test]
|
|
fn cli_convert_from_default_graph_to_named_graph() -> Result<()> {
|
|
cli_command()?
|
|
.arg("convert")
|
|
.arg("--from-format")
|
|
.arg("trig")
|
|
.arg("--to-format")
|
|
.arg("nq")
|
|
.arg("--from-default-graph")
|
|
.arg("--to-graph")
|
|
.arg("http://example.com/t")
|
|
.write_stdin("@base <http://example.com/> . <s> <p> <o> . <g> { <sg> <pg> <og> . }")
|
|
.assert()
|
|
.stdout("<http://example.com/s> <http://example.com/p> <http://example.com/o> <http://example.com/t> .\n")
|
|
.success();
|
|
Ok(())
|
|
}
|
|
|
|
#[test]
|
|
fn cli_convert_from_named_graph() -> Result<()> {
|
|
cli_command()?
|
|
.arg("convert")
|
|
.arg("--from-format")
|
|
.arg("trig")
|
|
.arg("--to-format")
|
|
.arg("nq")
|
|
.arg("--from-graph")
|
|
.arg("http://example.com/g")
|
|
.write_stdin("@base <http://example.com/> . <s> <p> <o> . <g> { <sg> <pg> <og> . }")
|
|
.assert()
|
|
.stdout("<http://example.com/sg> <http://example.com/pg> <http://example.com/og> .\n");
|
|
Ok(())
|
|
}
|
|
|
|
#[test]
|
|
fn get_ui() -> Result<()> {
|
|
ServerTest::new()?.test_status(
|
|
Request::builder(Method::GET, "http://localhost/".parse()?).build(),
|
|
Status::OK,
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn post_dataset_file() -> Result<()> {
|
|
let request = Request::builder(Method::POST, "http://localhost/store".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "application/trig")?
|
|
.with_body("<http://example.com> <http://example.com> <http://example.com> .");
|
|
ServerTest::new()?.test_status(request, Status::NO_CONTENT)
|
|
}
|
|
|
|
#[test]
|
|
fn post_wrong_file() -> Result<()> {
|
|
let request = Request::builder(Method::POST, "http://localhost/store".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "application/trig")?
|
|
.with_body("<http://example.com>");
|
|
ServerTest::new()?.test_status(request, Status::BAD_REQUEST)
|
|
}
|
|
|
|
#[test]
|
|
fn post_unsupported_file() -> Result<()> {
|
|
let request = Request::builder(Method::POST, "http://localhost/store".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/foo")?
|
|
.build();
|
|
ServerTest::new()?.test_status(request, Status::UNSUPPORTED_MEDIA_TYPE)
|
|
}
|
|
|
|
#[test]
|
|
fn get_query() -> Result<()> {
|
|
let server = ServerTest::new()?;
|
|
|
|
let request = Request::builder(Method::POST, "http://localhost/store".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "application/trig")?
|
|
.with_body("<http://example.com> <http://example.com> <http://example.com> .");
|
|
server.test_status(request, Status::NO_CONTENT)?;
|
|
|
|
let request = Request::builder(
|
|
Method::GET,
|
|
"http://localhost/query?query=SELECT%20?s%20?p%20?o%20WHERE%20{%20?s%20?p%20?o%20}"
|
|
.parse()?,
|
|
)
|
|
.with_header(HeaderName::ACCEPT, "text/csv")?
|
|
.build();
|
|
server.test_body(
|
|
request,
|
|
"s,p,o\r\nhttp://example.com,http://example.com,http://example.com\r\n",
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn get_query_accept_star() -> Result<()> {
|
|
let request = Request::builder(
|
|
Method::GET,
|
|
"http://localhost/query?query=SELECT%20?s%20?p%20?o%20WHERE%20{%20?s%20?p%20?o%20}"
|
|
.parse()?,
|
|
)
|
|
.with_header(HeaderName::ACCEPT, "*/*")?
|
|
.build();
|
|
ServerTest::new()?.test_body(
|
|
request,
|
|
r#"{"head":{"vars":["s","p","o"]},"results":{"bindings":[]}}"#,
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn get_query_accept_substar() -> Result<()> {
|
|
let request = Request::builder(
|
|
Method::GET,
|
|
"http://localhost/query?query=SELECT%20?s%20?p%20?o%20WHERE%20{%20?s%20?p%20?o%20}"
|
|
.parse()?,
|
|
)
|
|
.with_header(HeaderName::ACCEPT, "text/*")?
|
|
.build();
|
|
ServerTest::new()?.test_body(
|
|
request,
|
|
r#"{"head":{"vars":["s","p","o"]},"results":{"bindings":[]}}"#,
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn get_query_accept_good() -> Result<()> {
|
|
let request = Request::builder(
|
|
Method::GET,
|
|
"http://localhost/query?query=SELECT%20?s%20?p%20?o%20WHERE%20{%20?s%20?p%20?o%20}"
|
|
.parse()?,
|
|
)
|
|
.with_header(
|
|
HeaderName::ACCEPT,
|
|
"application/sparql-results+json;charset=utf-8",
|
|
)?
|
|
.build();
|
|
ServerTest::new()?.test_body(
|
|
request,
|
|
r#"{"head":{"vars":["s","p","o"]},"results":{"bindings":[]}}"#,
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn get_query_accept_bad() -> Result<()> {
|
|
let request = Request::builder(
|
|
Method::GET,
|
|
"http://localhost/query?query=SELECT%20?s%20?p%20?o%20WHERE%20{%20?s%20?p%20?o%20}"
|
|
.parse()?,
|
|
)
|
|
.with_header(HeaderName::ACCEPT, "application/foo")?
|
|
.build();
|
|
ServerTest::new()?.test_status(request, Status::NOT_ACCEPTABLE)
|
|
}
|
|
|
|
#[test]
|
|
fn get_query_accept_explicit_priority() -> Result<()> {
|
|
let request = Request::builder(
|
|
Method::GET,
|
|
"http://localhost/query?query=SELECT%20?s%20?p%20?o%20WHERE%20{%20?s%20?p%20?o%20}"
|
|
.parse()?,
|
|
)
|
|
.with_header(HeaderName::ACCEPT, "text/foo;q=0.5 , text/json ; q = 0.7")?
|
|
.build();
|
|
ServerTest::new()?.test_body(
|
|
request,
|
|
r#"{"head":{"vars":["s","p","o"]},"results":{"bindings":[]}}"#,
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn get_query_accept_implicit_priority() -> Result<()> {
|
|
let request = Request::builder(
|
|
Method::GET,
|
|
"http://localhost/query?query=SELECT%20?s%20?p%20?o%20WHERE%20{%20?s%20?p%20?o%20}"
|
|
.parse()?,
|
|
)
|
|
.with_header(HeaderName::ACCEPT, "text/json,text/foo")?
|
|
.build();
|
|
ServerTest::new()?.test_body(
|
|
request,
|
|
r#"{"head":{"vars":["s","p","o"]},"results":{"bindings":[]}}"#,
|
|
)
|
|
}
|
|
#[test]
|
|
fn get_query_accept_implicit_and_explicit_priority() -> Result<()> {
|
|
let request = Request::builder(
|
|
Method::GET,
|
|
"http://localhost/query?query=SELECT%20?s%20?p%20?o%20WHERE%20{%20?s%20?p%20?o%20}"
|
|
.parse()?,
|
|
)
|
|
.with_header(HeaderName::ACCEPT, "text/foo;q=0.9,text/csv")?
|
|
.build();
|
|
ServerTest::new()?.test_body(request, "s,p,o\r\n")
|
|
}
|
|
|
|
#[test]
|
|
fn get_bad_query() -> Result<()> {
|
|
ServerTest::new()?.test_status(
|
|
Request::builder(Method::GET, "http://localhost/query?query=SELECT".parse()?).build(),
|
|
Status::BAD_REQUEST,
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn get_query_union_graph() -> Result<()> {
|
|
let server = ServerTest::new()?;
|
|
|
|
let request = Request::builder(Method::PUT, "http://localhost/store/1".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/turtle")?
|
|
.with_body("<http://example.com> <http://example.com> <http://example.com> .");
|
|
server.test_status(request, Status::CREATED)?;
|
|
|
|
let request = Request::builder(
|
|
Method::GET,
|
|
"http://localhost/query?query=SELECT%20?s%20?p%20?o%20WHERE%20{%20?s%20?p%20?o%20}&union-default-graph"
|
|
.parse()
|
|
?,
|
|
).with_header(HeaderName::ACCEPT, "text/csv")
|
|
?
|
|
.build();
|
|
server.test_body(
|
|
request,
|
|
"s,p,o\r\nhttp://example.com,http://example.com,http://example.com\r\n",
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn get_query_union_graph_in_url_and_urlencoded() -> Result<()> {
|
|
let server = ServerTest::new()?;
|
|
|
|
let request = Request::builder(Method::PUT, "http://localhost/store/1".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/turtle")?
|
|
.with_body("<http://example.com> <http://example.com> <http://example.com> .");
|
|
server.test_status(request, Status::CREATED)?;
|
|
|
|
let request = Request::builder(
|
|
Method::POST,
|
|
"http://localhost/query?union-default-graph".parse()?,
|
|
)
|
|
.with_header(
|
|
HeaderName::CONTENT_TYPE,
|
|
"application/x-www-form-urlencoded",
|
|
)?
|
|
.with_header(HeaderName::ACCEPT, "text/csv")?
|
|
.with_body("query=SELECT%20?s%20?p%20?o%20WHERE%20{%20?s%20?p%20?o%20}");
|
|
server.test_body(
|
|
request,
|
|
"s,p,o\r\nhttp://example.com,http://example.com,http://example.com\r\n",
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn get_query_union_graph_and_default_graph() -> Result<()> {
|
|
ServerTest::new()?.test_status(Request::builder(
|
|
Method::GET,
|
|
"http://localhost/query?query=SELECT%20*%20WHERE%20{%20?s%20?p%20?o%20}&union-default-graph&default-graph-uri=http://example.com".parse()
|
|
?,
|
|
).build(), Status::BAD_REQUEST)
|
|
}
|
|
|
|
#[test]
|
|
fn get_without_query() -> Result<()> {
|
|
ServerTest::new()?.test_status(
|
|
Request::builder(Method::GET, "http://localhost/query".parse()?).build(),
|
|
Status::BAD_REQUEST,
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn post_query() -> Result<()> {
|
|
let request = Request::builder(Method::POST, "http://localhost/query".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "application/sparql-query")?
|
|
.with_body("SELECT * WHERE { ?s ?p ?o }");
|
|
ServerTest::new()?.test_status(request, Status::OK)
|
|
}
|
|
|
|
#[test]
|
|
fn post_bad_query() -> Result<()> {
|
|
let request = Request::builder(Method::POST, "http://localhost/query".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "application/sparql-query")?
|
|
.with_body("SELECT");
|
|
ServerTest::new()?.test_status(request, Status::BAD_REQUEST)
|
|
}
|
|
|
|
#[test]
|
|
fn post_unknown_query() -> Result<()> {
|
|
let request = Request::builder(Method::POST, "http://localhost/query".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "application/sparql-todo")?
|
|
.with_body("SELECT");
|
|
ServerTest::new()?.test_status(request, Status::UNSUPPORTED_MEDIA_TYPE)
|
|
}
|
|
|
|
#[test]
|
|
fn post_federated_query_wikidata() -> Result<()> {
|
|
let request = Request::builder(Method::POST, "http://localhost/query".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "application/sparql-query")
|
|
?.with_body("SELECT * WHERE { SERVICE <https://query.wikidata.org/sparql> { <https://en.wikipedia.org/wiki/Paris> ?p ?o } }");
|
|
ServerTest::new()?.test_status(request, Status::OK)
|
|
}
|
|
|
|
#[test]
|
|
fn post_federated_query_dbpedia() -> Result<()> {
|
|
let request = Request::builder(Method::POST, "http://localhost/query".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "application/sparql-query")
|
|
?.with_body("SELECT * WHERE { SERVICE <https://dbpedia.org/sparql> { <http://dbpedia.org/resource/Paris> ?p ?o } }");
|
|
ServerTest::new()?.test_status(request, Status::OK)
|
|
}
|
|
|
|
#[test]
|
|
fn post_update() -> Result<()> {
|
|
let request = Request::builder(Method::POST, "http://localhost/update".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "application/sparql-update")?
|
|
.with_body(
|
|
"INSERT DATA { <http://example.com> <http://example.com> <http://example.com> }",
|
|
);
|
|
ServerTest::new()?.test_status(request, Status::NO_CONTENT)
|
|
}
|
|
|
|
#[test]
|
|
fn post_bad_update() -> Result<()> {
|
|
let request = Request::builder(Method::POST, "http://localhost/update".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "application/sparql-update")?
|
|
.with_body("INSERT");
|
|
ServerTest::new()?.test_status(request, Status::BAD_REQUEST)
|
|
}
|
|
|
|
#[test]
|
|
fn post_update_read_only() -> Result<()> {
|
|
let request = Request::builder(Method::POST, "http://localhost/update".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "application/sparql-update")?
|
|
.with_body(
|
|
"INSERT DATA { <http://example.com> <http://example.com> <http://example.com> }",
|
|
);
|
|
ServerTest::check_status(
|
|
ServerTest::new()?.exec_read_only(request),
|
|
Status::FORBIDDEN,
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn graph_store_url_normalization() -> Result<()> {
|
|
let server = ServerTest::new()?;
|
|
|
|
// PUT
|
|
let request = Request::builder(
|
|
Method::PUT,
|
|
"http://localhost/store?graph=http://example.com".parse()?,
|
|
)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/turtle")?
|
|
.with_body("<http://example.com> <http://example.com> <http://example.com> .");
|
|
server.test_status(request, Status::CREATED)?;
|
|
|
|
// GET good URI
|
|
server.test_status(
|
|
Request::builder(
|
|
Method::GET,
|
|
"http://localhost/store?graph=http://example.com".parse()?,
|
|
)
|
|
.build(),
|
|
Status::OK,
|
|
)?;
|
|
|
|
// GET bad URI
|
|
server.test_status(
|
|
Request::builder(
|
|
Method::GET,
|
|
"http://localhost/store?graph=http://example.com/".parse()?,
|
|
)
|
|
.build(),
|
|
Status::NOT_FOUND,
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn graph_store_base_url() -> Result<()> {
|
|
let server = ServerTest::new()?;
|
|
|
|
// POST
|
|
let request = Request::builder(
|
|
Method::POST,
|
|
"http://localhost/store?graph=http://example.com".parse()?,
|
|
)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/turtle")?
|
|
.with_body("<> <http://example.com/p> <http://example.com/o1> .");
|
|
server.test_status(request, Status::NO_CONTENT)?;
|
|
|
|
// GET
|
|
let request = Request::builder(
|
|
Method::GET,
|
|
"http://localhost/store?graph=http://example.com".parse()?,
|
|
)
|
|
.with_header(HeaderName::ACCEPT, "application/n-triples")?
|
|
.build();
|
|
server.test_body(
|
|
request,
|
|
"<http://example.com> <http://example.com/p> <http://example.com/o1> .\n",
|
|
)?;
|
|
|
|
// PUT
|
|
let request = Request::builder(
|
|
Method::PUT,
|
|
"http://localhost/store?graph=http://example.com".parse()?,
|
|
)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/turtle")?
|
|
.with_body("<> <http://example.com/p> <http://example.com/o2> .");
|
|
server.test_status(request, Status::NO_CONTENT)?;
|
|
|
|
// GET
|
|
let request = Request::builder(
|
|
Method::GET,
|
|
"http://localhost/store?graph=http://example.com".parse()?,
|
|
)
|
|
.with_header(HeaderName::ACCEPT, "application/n-triples")?
|
|
.build();
|
|
server.test_body(
|
|
request,
|
|
"<http://example.com> <http://example.com/p> <http://example.com/o2> .\n",
|
|
)
|
|
}
|
|
|
|
#[test]
|
|
fn graph_store_protocol() -> Result<()> {
|
|
// Tests from https://www.w3.org/2009/sparql/docs/tests/data-sparql11/http-rdf-update/
|
|
|
|
let server = ServerTest::new()?;
|
|
|
|
// PUT - Initial state
|
|
let request = Request::builder(Method::PUT, "http://localhost/store/person/1.ttl".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/turtle; charset=utf-8")?
|
|
.with_body(
|
|
r#"
|
|
@prefix foaf: <http://xmlns.com/foaf/0.1/> .
|
|
@prefix v: <http://www.w3.org/2006/vcard/ns#> .
|
|
|
|
<http://$HOST$/$GRAPHSTORE$/person/1> a foaf:Person;
|
|
foaf:businessCard [
|
|
a v:VCard;
|
|
v:fn "John Doe"
|
|
].
|
|
"#,
|
|
);
|
|
server.test_status(request, Status::CREATED)?;
|
|
|
|
// GET of PUT - Initial state
|
|
let request = Request::builder(
|
|
Method::GET,
|
|
"http://localhost/store?graph=/store/person/1.ttl".parse()?,
|
|
)
|
|
.with_header(HeaderName::ACCEPT, "text/turtle")?
|
|
.build();
|
|
server.test_status(request, Status::OK)?;
|
|
|
|
// HEAD on an existing graph
|
|
server.test_status(
|
|
Request::builder(Method::HEAD, "http://localhost/store/person/1.ttl".parse()?).build(),
|
|
Status::OK,
|
|
)?;
|
|
|
|
// HEAD on a non-existing graph
|
|
server.test_status(
|
|
Request::builder(Method::HEAD, "http://localhost/store/person/4.ttl".parse()?).build(),
|
|
Status::NOT_FOUND,
|
|
)?;
|
|
|
|
// PUT - graph already in store
|
|
let request = Request::builder(Method::PUT, "http://localhost/store/person/1.ttl".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/turtle; charset=utf-8")?
|
|
.with_body(
|
|
r#"
|
|
@prefix foaf: <http://xmlns.com/foaf/0.1/> .
|
|
@prefix v: <http://www.w3.org/2006/vcard/ns#> .
|
|
|
|
<http://$HOST$/$GRAPHSTORE$/person/1> a foaf:Person;
|
|
foaf:businessCard [
|
|
a v:VCard;
|
|
v:fn "Jane Doe"
|
|
].
|
|
"#,
|
|
);
|
|
server.test_status(request, Status::NO_CONTENT)?;
|
|
|
|
// GET of PUT - graph already in store
|
|
let request = Request::builder(Method::GET, "http://localhost/store/person/1.ttl".parse()?)
|
|
.with_header(HeaderName::ACCEPT, "text/turtle")?
|
|
.build();
|
|
server.test_status(request, Status::OK)?;
|
|
|
|
// PUT - default graph
|
|
let request = Request::builder(Method::PUT, "http://localhost/store?default".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/turtle; charset=utf-8")?
|
|
.with_body(
|
|
r#"
|
|
@prefix foaf: <http://xmlns.com/foaf/0.1/> .
|
|
@prefix v: <http://www.w3.org/2006/vcard/ns#> .
|
|
|
|
[] a foaf:Person;
|
|
foaf:businessCard [
|
|
a v:VCard;
|
|
v:given-name "Alice"
|
|
] .
|
|
"#,
|
|
);
|
|
server.test_status(request, Status::NO_CONTENT)?; // The default graph always exists in Oxigraph
|
|
|
|
// GET of PUT - default graph
|
|
let request = Request::builder(Method::GET, "http://localhost/store?default".parse()?)
|
|
.with_header(HeaderName::ACCEPT, "text/turtle")?
|
|
.build();
|
|
server.test_status(request, Status::OK)?;
|
|
|
|
// PUT - mismatched payload
|
|
let request = Request::builder(Method::PUT, "http://localhost/store/person/1.ttl".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/turtle; charset=utf-8")?
|
|
.with_body("@prefix foo");
|
|
server.test_status(request, Status::BAD_REQUEST)?;
|
|
|
|
// PUT - empty graph
|
|
let request = Request::builder(Method::PUT, "http://localhost/store/person/2.ttl".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/turtle; charset=utf-8")?
|
|
.build();
|
|
server.test_status(request, Status::CREATED)?;
|
|
|
|
// GET of PUT - empty graph
|
|
let request = Request::builder(Method::GET, "http://localhost/store/person/2.ttl".parse()?)
|
|
.with_header(HeaderName::ACCEPT, "text/turtle")?
|
|
.build();
|
|
server.test_status(request, Status::OK)?;
|
|
|
|
// PUT - replace empty graph
|
|
let request = Request::builder(Method::PUT, "http://localhost/store/person/2.ttl".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/turtle; charset=utf-8")?
|
|
.with_body(
|
|
r#"
|
|
@prefix foaf: <http://xmlns.com/foaf/0.1/> .
|
|
@prefix v: <http://www.w3.org/2006/vcard/ns#> .
|
|
|
|
[] a foaf:Person;
|
|
foaf:businessCard [
|
|
a v:VCard;
|
|
v:given-name "Alice"
|
|
] .
|
|
"#,
|
|
);
|
|
server.test_status(request, Status::NO_CONTENT)?;
|
|
|
|
// GET of replacement for empty graph
|
|
let request = Request::builder(Method::GET, "http://localhost/store/person/2.ttl".parse()?)
|
|
.with_header(HeaderName::ACCEPT, "text/turtle")?
|
|
.build();
|
|
server.test_status(request, Status::OK)?;
|
|
|
|
// DELETE - existing graph
|
|
server.test_status(
|
|
Request::builder(
|
|
Method::DELETE,
|
|
"http://localhost/store/person/2.ttl".parse()?,
|
|
)
|
|
.build(),
|
|
Status::NO_CONTENT,
|
|
)?;
|
|
|
|
// GET of DELETE - existing graph
|
|
server.test_status(
|
|
Request::builder(Method::GET, "http://localhost/store/person/2.ttl".parse()?).build(),
|
|
Status::NOT_FOUND,
|
|
)?;
|
|
|
|
// DELETE - non-existent graph
|
|
server.test_status(
|
|
Request::builder(
|
|
Method::DELETE,
|
|
"http://localhost/store/person/2.ttl".parse()?,
|
|
)
|
|
.build(),
|
|
Status::NOT_FOUND,
|
|
)?;
|
|
|
|
// POST - existing graph
|
|
let request = Request::builder(Method::PUT, "http://localhost/store/person/1.ttl".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/turtle; charset=utf-8")?
|
|
.build();
|
|
server.test_status(request, Status::NO_CONTENT)?;
|
|
|
|
// TODO: POST - multipart/form-data
|
|
// TODO: GET of POST - multipart/form-data
|
|
|
|
// POST - create new graph
|
|
let request = Request::builder(Method::POST, "http://localhost/store".parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/turtle; charset=utf-8")?
|
|
.with_body(
|
|
r#"
|
|
@prefix foaf: <http://xmlns.com/foaf/0.1/> .
|
|
@prefix v: <http://www.w3.org/2006/vcard/ns#> .
|
|
|
|
[] a foaf:Person;
|
|
foaf:businessCard [
|
|
a v:VCard;
|
|
v:given-name "Alice"
|
|
] .
|
|
"#,
|
|
);
|
|
let response = server.exec(request);
|
|
assert_eq!(response.status(), Status::CREATED);
|
|
let location = response.header(&HeaderName::LOCATION).unwrap().to_str()?;
|
|
|
|
// GET of POST - create new graph
|
|
let request = Request::builder(Method::GET, location.parse()?)
|
|
.with_header(HeaderName::ACCEPT, "text/turtle")?
|
|
.build();
|
|
server.test_status(request, Status::OK)?;
|
|
|
|
// POST - empty graph to existing graph
|
|
let request = Request::builder(Method::PUT, location.parse()?)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/turtle; charset=utf-8")?
|
|
.build();
|
|
server.test_status(request, Status::NO_CONTENT)?;
|
|
|
|
// GET of POST - after noop
|
|
let request = Request::builder(Method::GET, location.parse()?)
|
|
.with_header(HeaderName::ACCEPT, "text/turtle")?
|
|
.build();
|
|
server.test_status(request, Status::OK)
|
|
}
|
|
|
|
#[test]
|
|
fn graph_store_lenient_bulk() -> Result<()> {
|
|
let server = ServerTest::new()?;
|
|
let invalid_data = "
|
|
@prefix foaf: <http://xmlns.com/foaf/0.1/> .
|
|
@prefix v: <http://www.w3.org/2006/vcard/ns#> .
|
|
|
|
<http://$HOST$/$GRAPHSTORE$/person/1> a foaf:Person . foo";
|
|
|
|
// POST
|
|
let request = Request::builder(
|
|
Method::POST,
|
|
"http://localhost/store/person/1.ttl?no_transaction&lenient".parse()?,
|
|
)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/turtle; charset=utf-8")?
|
|
.with_body(invalid_data);
|
|
server.test_status(request, Status::NO_CONTENT)?;
|
|
|
|
// GET of POST
|
|
let request = Request::builder(
|
|
Method::GET,
|
|
"http://localhost/store?graph=/store/person/1.ttl".parse()?,
|
|
)
|
|
.with_header(HeaderName::ACCEPT, "text/turtle")?
|
|
.build();
|
|
server.test_status(request, Status::OK)?;
|
|
|
|
// POST dataset
|
|
let request = Request::builder(
|
|
Method::POST,
|
|
"http://localhost/store?lenient&no_transaction".parse()?,
|
|
)
|
|
.with_header(HeaderName::CONTENT_TYPE, "application/trig; charset=utf-8")?
|
|
.with_body(invalid_data);
|
|
server.test_status(request, Status::NO_CONTENT)?;
|
|
|
|
// GET of POST dataset
|
|
let request = Request::builder(Method::GET, "http://localhost/store?default".parse()?)
|
|
.with_header(HeaderName::ACCEPT, "text/turtle")?
|
|
.build();
|
|
server.test_status(request, Status::OK)?;
|
|
|
|
// PUT
|
|
let request = Request::builder(
|
|
Method::PUT,
|
|
"http://localhost/store/person/1.ttl?lenient&no_transaction".parse()?,
|
|
)
|
|
.with_header(HeaderName::CONTENT_TYPE, "text/turtle; charset=utf-8")?
|
|
.with_body(invalid_data);
|
|
server.test_status(request, Status::NO_CONTENT)?;
|
|
|
|
// GET of PUT - Initial state
|
|
let request = Request::builder(
|
|
Method::GET,
|
|
"http://localhost/store?graph=/store/person/1.ttl".parse()?,
|
|
)
|
|
.with_header(HeaderName::ACCEPT, "text/turtle")?
|
|
.build();
|
|
server.test_status(request, Status::OK)?;
|
|
|
|
// PUT dataset
|
|
let request = Request::builder(
|
|
Method::PUT,
|
|
"http://localhost/store?lenient&no_transaction".parse()?,
|
|
)
|
|
.with_header(HeaderName::CONTENT_TYPE, "application/trig; charset=utf-8")?
|
|
.with_body(invalid_data);
|
|
server.test_status(request, Status::NO_CONTENT)?;
|
|
|
|
// GET of PUT dataset
|
|
let request = Request::builder(Method::GET, "http://localhost/store?default".parse()?)
|
|
.with_header(HeaderName::ACCEPT, "text/turtle")?
|
|
.build();
|
|
server.test_status(request, Status::OK)?;
|
|
|
|
// GET of PUT dataset - replacement
|
|
let request = Request::builder(
|
|
Method::GET,
|
|
"http://localhost/store?graph=/store/person/1.ttl".parse()?,
|
|
)
|
|
.with_header(HeaderName::ACCEPT, "text/turtle")?
|
|
.build();
|
|
server.test_status(request, Status::NOT_FOUND)
|
|
}
|
|
|
|
struct ServerTest {
|
|
store: Store,
|
|
}
|
|
|
|
impl ServerTest {
|
|
fn new() -> Result<Self> {
|
|
Ok(Self {
|
|
store: Store::new()?,
|
|
})
|
|
}
|
|
|
|
fn exec(&self, mut request: Request) -> Response {
|
|
handle_request(&mut request, self.store.clone(), false)
|
|
.unwrap_or_else(|(status, message)| error(status, message))
|
|
}
|
|
|
|
fn exec_read_only(&self, mut request: Request) -> Response {
|
|
handle_request(&mut request, self.store.clone(), true)
|
|
.unwrap_or_else(|(status, message)| error(status, message))
|
|
}
|
|
|
|
fn test_status(&self, request: Request, expected_status: Status) -> Result<()> {
|
|
Self::check_status(self.exec(request), expected_status)
|
|
}
|
|
|
|
fn check_status(mut response: Response, expected_status: Status) -> Result<()> {
|
|
let body = read_to_string(response.body_mut())?;
|
|
assert_eq!(response.status(), expected_status, "Error message: {body}");
|
|
Ok(())
|
|
}
|
|
|
|
fn test_body(&self, request: Request, expected_body: &str) -> Result<()> {
|
|
let mut response = self.exec(request);
|
|
let body = read_to_string(response.body_mut())?;
|
|
assert_eq!(response.status(), Status::OK, "Error message: {body}");
|
|
assert_eq!(&body, expected_body);
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
#[test]
|
|
fn clap_debug() {
|
|
use clap::CommandFactory;
|
|
|
|
Args::command().debug_assert()
|
|
}
|
|
}
|
|
|