/*
* Licensed to the Apache Software Foundation (ASF) under one
* or more contributor license agreements. See the NOTICE file
* distributed with this work for additional information
* regarding copyright ownership. The ASF licenses this file
* to you under the Apache License, Version 2.0 (the
* "License"); you may not use this file except in compliance
* with the License. You may obtain a copy of the License at
*
* http://www.apache.org/licenses/LICENSE-2.0
*
* Unless required by applicable law or agreed to in writing, software
* distributed under the License is distributed on an "AS IS" BASIS,
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
* See the License for the specific language governing permissions and
* limitations under the License.
*/
package com.hp.hpl.jena.tdb.sys ;
import static com.hp.hpl.jena.tdb.sys.SystemTDB.LenNodeHash ;
import static com.hp.hpl.jena.tdb.sys.SystemTDB.SizeOfNodeId ;
import org.openjena.atlas.lib.ColumnMap ;
import org.openjena.atlas.lib.StrUtils ;
import org.slf4j.Logger ;
import com.hp.hpl.jena.tdb.TDB ;
import com.hp.hpl.jena.tdb.TDBException ;
import com.hp.hpl.jena.tdb.base.block.BlockMgr ;
import com.hp.hpl.jena.tdb.base.block.BlockMgrFactory ;
import com.hp.hpl.jena.tdb.base.file.FileFactory ;
import com.hp.hpl.jena.tdb.base.file.FileSet ;
import com.hp.hpl.jena.tdb.base.file.Location ;
import com.hp.hpl.jena.tdb.base.objectfile.ObjectFile ;
import com.hp.hpl.jena.tdb.base.record.RecordFactory ;
import com.hp.hpl.jena.tdb.index.Index ;
import com.hp.hpl.jena.tdb.index.IndexBuilder ;
import com.hp.hpl.jena.tdb.index.RangeIndex ;
import com.hp.hpl.jena.tdb.index.TupleIndex ;
import com.hp.hpl.jena.tdb.index.TupleIndexRecord ;
import com.hp.hpl.jena.tdb.index.bplustree.BPlusTree ;
import com.hp.hpl.jena.tdb.index.bplustree.BPlusTreeParams ;
import com.hp.hpl.jena.tdb.nodetable.NodeTable ;
import com.hp.hpl.jena.tdb.nodetable.NodeTableCache ;
import com.hp.hpl.jena.tdb.nodetable.NodeTableFactory ;
import com.hp.hpl.jena.tdb.nodetable.NodeTableInline ;
import com.hp.hpl.jena.tdb.nodetable.NodeTableNative ;
import com.hp.hpl.jena.tdb.setup.DatasetBuilderStd ;
import com.hp.hpl.jena.tdb.setup.SystemParams ;
import com.hp.hpl.jena.tdb.store.DatasetGraphTDB ;
import com.hp.hpl.jena.tdb.store.DatasetPrefixesTDB ;
import com.hp.hpl.jena.tdb.store.NodeId ;
import com.hp.hpl.jena.tdb.store.QuadTable ;
import com.hp.hpl.jena.tdb.store.TripleTable ;
/** Makes things : datasets from locations, indexes, etc etc. */
public class SetupTDB
{
// Replaced/merge with DatasetBuilderStd mechanisms.
//private static final Logger log = LoggerFactory.getLogger(NewSetup.class) ;
static final Logger log = TDB.logInfo ;
public static void error(Logger log, String msg)
{
if ( log != null )
log.error(msg) ;
throw new TDBException(msg) ;
}
private static SystemParams params = SystemParams.getStdSystemParams() ;
// And here we make datasets ...
public static DatasetGraphTDB buildDataset(Location location)
{
return DatasetBuilderStd.build(location) ;
}
//protected static DatasetControl createConcurrencyPolicy() { return new DatasetControlMRSW() ; }
public static TripleTable makeTripleTable(Location location, NodeTable nodeTable, String dftPrimary, String[] dftIndexes, DatasetControl policy)
{
String primary = Names.primaryIndexTriples ;
String indexes[] = Names.tripleIndexes ;
if ( indexes.length != 3 )
error(log, "Wrong number of triple table indexes: "+StrUtils.strjoin(",", indexes)) ;
log.debug("Triple table: "+primary+" :: "+StrUtils.strjoin(",", indexes)) ;
TupleIndex tripleIndexes[] = makeTupleIndexes(location, primary, indexes, indexes) ;
if ( tripleIndexes.length != indexes.length )
error(log, "Wrong number of triple table tuples indexes: "+tripleIndexes.length) ;
TripleTable tripleTable = new TripleTable(tripleIndexes, nodeTable, policy) ;
return tripleTable ;
}
public static QuadTable makeQuadTable(Location location, NodeTable nodeTable, String dftPrimary, String[] dftIndexes, DatasetControl policy)
{
String primary = Names.primaryIndexQuads ;
String indexes[] = Names.quadIndexes ;
if ( indexes.length != 6 )
error(log, "Wrong number of quad table indexes: "+StrUtils.strjoin(",", indexes)) ;
log.debug("Quad table: "+primary+" :: "+StrUtils.strjoin(",", indexes)) ;
TupleIndex quadIndexes[] = makeTupleIndexes(location, primary, indexes, indexes) ;
if ( quadIndexes.length != indexes.length )
error(log, "Wrong number of quad table tuples indexes: "+quadIndexes.length) ;
QuadTable quadTable = new QuadTable(quadIndexes, nodeTable, policy) ;
return quadTable ;
}
public static DatasetPrefixesTDB makePrefixes(Location location, DatasetControl policy)
{
// The index using for Graph+Prefix => URI
String indexPrefixes = params.indexPrefix ;
String primary = params.primaryIndexPrefix ;
String indexes[] = params.prefixIndexes ;
TupleIndex prefixIndexes[] = makeTupleIndexes(location, primary, indexes, new String[]{indexPrefixes}) ;
if ( prefixIndexes.length != indexes.length )
error(log, "Wrong number of triple table tuples indexes: "+prefixIndexes.length) ;
// The nodetable.
String pnNode2Id = params.prefixNode2Id ;
String pnId2Node = params.prefixId2Node ;
// No cache - the prefix mapping is a cache
NodeTable prefixNodes = makeNodeTable(location, pnNode2Id, -1, pnId2Node, -1, -1) ;
DatasetPrefixesTDB prefixes = new DatasetPrefixesTDB(prefixIndexes, prefixNodes, policy) ;
log.debug("Prefixes: "+StrUtils.strjoin(", ", indexes)) ;
return prefixes ;
}
public static TupleIndex[] makeTupleIndexes(Location location, String primary, String[] descs, String[] filenames)
{
if ( primary.length() != 3 && primary.length() != 4 )
error(log, "Bad primary key length: "+primary.length()) ;
int indexRecordLen = primary.length()*NodeId.SIZE ;
TupleIndex indexes[] = new TupleIndex[descs.length] ;
for (int i = 0 ; i < indexes.length ; i++)
indexes[i] = makeTupleIndex(location, primary, descs[i], filenames[i], indexRecordLen) ;
return indexes ;
}
public static TupleIndex makeTupleIndex(Location location,
String primary, String indexOrder, String indexName,
int keyLength)
{
FileSet fs = new FileSet(location, indexName) ;
int readCacheSize = params.readCacheSize ;
int writeCacheSize = params.writeCacheSize ;
// Value part is null (zero length)
RangeIndex rIndex = makeRangeIndex(location, indexName, keyLength, 0, readCacheSize, writeCacheSize) ;
TupleIndex tupleIndex = new TupleIndexRecord(primary.length(), new ColumnMap(primary, indexOrder), rIndex.getRecordFactory(), rIndex) ;
return tupleIndex ;
}
public static Index makeIndex(Location location, String indexName,
int dftKeyLength, int dftValueLength,
int readCacheSize,int writeCacheSize)
{
return makeRangeIndex(location, indexName, dftKeyLength, dftValueLength, readCacheSize, writeCacheSize) ;
}
public static RangeIndex makeRangeIndex(Location location, String indexName,
int dftKeyLength, int dftValueLength,
int readCacheSize,int writeCacheSize)
{
FileSet fs = new FileSet(location, indexName) ;
RangeIndex rIndex = makeBPlusTree(fs, readCacheSize, writeCacheSize, dftKeyLength, dftValueLength) ;
return rIndex ;
}
public static RangeIndex makeBPlusTree(FileSet fs,
int readCacheSize, int writeCacheSize,
int dftKeyLength, int dftValueLength)
{
RecordFactory recordFactory = makeRecordFactory(dftKeyLength, dftValueLength) ;
int blkSize = params.blockSize ;
// IndexBuilder.getBPlusTree().newRangeIndex(fs, recordFactory) ;
// Does not set order.
int order = BPlusTreeParams.calcOrder(blkSize, recordFactory.recordLength()) ;
RangeIndex rIndex = createBPTree(fs, order, blkSize, readCacheSize, writeCacheSize, recordFactory) ;
return rIndex ;
}
public static RecordFactory makeRecordFactory(int keyLen, int valueLen)
{
return new RecordFactory(keyLen, valueLen) ;
}
/** Make a NodeTable without cache and inline wrappers */
public static NodeTable makeNodeTableBase(Location location, String indexNode2Id, String indexId2Node)
{
if (location.isMem())
return NodeTableFactory.createMem(IndexBuilder.mem()) ;
// -- make id to node mapping -- Names.indexId2Node
FileSet fsIdToNode = new FileSet(location, indexId2Node) ;
ObjectFile stringFile = makeObjectFile(fsIdToNode) ;
// -- make node to id mapping -- Names.indexNode2Id
// Make index of id to node (data table)
// No caching at the index level - we use the internal caches of the node table.
Index nodeToId = makeIndex(location, indexNode2Id, LenNodeHash, SizeOfNodeId, -1 ,-1) ;
// -- Make the node table using the components established above.
NodeTable nodeTable = new NodeTableNative(nodeToId, stringFile) ;
return nodeTable ;
}
/** Make a NodeTable with cache and inline wrappers */
public static NodeTable makeNodeTable(Location location)
{
return makeNodeTable(location,
Names.indexNode2Id, SystemTDB.Node2NodeIdCacheSize,
Names.indexId2Node, SystemTDB.NodeId2NodeCacheSize,
SystemTDB.NodeMissCacheSize) ;
}
/** Make a NodeTable with cache and inline wrappers */
public static NodeTable makeNodeTable(Location location,
String indexNode2Id, int nodeToIdCacheSize,
String indexId2Node, int idToNodeCacheSize,
int nodeMissCacheSize)
{
NodeTable nodeTable = makeNodeTableBase(location, indexNode2Id, indexId2Node) ;
nodeTable = NodeTableCache.create(nodeTable, nodeToIdCacheSize, idToNodeCacheSize, nodeMissCacheSize) ;
nodeTable = NodeTableInline.create(nodeTable) ;
return nodeTable ;
}
public static ObjectFile makeObjectFile(FileSet fsIdToNode)
{
String filename = fsIdToNode.filename(Names.extNodeData) ;
ObjectFile objFile = FileFactory.createObjectFileDisk(filename);
return objFile ;
}
/** Create a B+Tree using defaults */
public static RangeIndex createBPTree(FileSet fileset,
RecordFactory factory)
{
int readCacheSize = SystemTDB.BlockReadCacheSize ;
int writeCacheSize = SystemTDB.BlockWriteCacheSize ;
int blockSize = SystemTDB.BlockSize ;
if ( fileset.isMem() )
{
readCacheSize = 0 ;
writeCacheSize = 0 ;
blockSize = SystemTDB.BlockSizeTest ;
}
return createBPTreeByBlockSize(fileset, blockSize, readCacheSize, writeCacheSize, factory) ;
}
/** Create a B+Tree by BlockSize */
public static RangeIndex createBPTreeByBlockSize(FileSet fileset,
int blockSize,
int readCacheSize, int writeCacheSize,
RecordFactory factory)
{
return createBPTree(fileset, -1, blockSize, readCacheSize, writeCacheSize, factory) ;
}
/** Create a B+Tree by Order */
public static RangeIndex createBPTreeByOrder(FileSet fileset,
int order,
int readCacheSize, int writeCacheSize,
RecordFactory factory)
{
return createBPTree(fileset, order, -1, readCacheSize, writeCacheSize, factory) ;
}
/** Knowing all the parameters, create a B+Tree */
public static BPlusTree createBPTree(FileSet fileset, int order, int blockSize,
int readCacheSize, int writeCacheSize,
RecordFactory factory)
{
// ---- Checking
if (blockSize < 0 && order < 0) throw new IllegalArgumentException("Neither blocksize nor order specified") ;
if (blockSize >= 0 && order < 0) order = BPlusTreeParams.calcOrder(blockSize, factory.recordLength()) ;
if (blockSize >= 0 && order >= 0)
{
int order2 = BPlusTreeParams.calcOrder(blockSize, factory.recordLength()) ;
if (order != order2) throw new IllegalArgumentException("Wrong order (" + order + "), calculated = "
+ order2) ;
}
// Iffy - does not allow for slop.
if (blockSize < 0 && order >= 0)
{
// Only in-memory.
blockSize = BPlusTreeParams.calcBlockSize(order, factory) ;
}
BPlusTreeParams params = new BPlusTreeParams(order, factory) ;
BlockMgr blkMgrNodes = BlockMgrFactory.create(fileset, Names.bptExtTree, blockSize, readCacheSize, writeCacheSize) ;
BlockMgr blkMgrRecords = BlockMgrFactory.create(fileset, Names.bptExtRecords, blockSize, readCacheSize, writeCacheSize) ;
return BPlusTree.create(params, blkMgrNodes, blkMgrRecords) ;
}
}