Remove unused methods in DatasetUtil, IndexUtil, and MetadataProvider.
Change-Id: I60352d052f684a65fd823a1c7aa33b794bd7f476
Reviewed-on: https://asterix-gerrit.ics.uci.edu/1741
Sonar-Qube: Jenkins <jenkins@fulliautomatix.ics.uci.edu>
Reviewed-by: Till Westmann <tillw@apache.org>
Integration-Tests: Jenkins <jenkins@fulliautomatix.ics.uci.edu>
Tested-by: Jenkins <jenkins@fulliautomatix.ics.uci.edu>
diff --git a/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/declared/MetadataProvider.java b/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/declared/MetadataProvider.java
index 5934f5e..750f89b 100644
--- a/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/declared/MetadataProvider.java
+++ b/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/declared/MetadataProvider.java
@@ -35,7 +35,6 @@
import org.apache.asterix.common.dataflow.LSMTreeInsertDeleteOperatorDescriptor;
import org.apache.asterix.common.exceptions.AsterixException;
import org.apache.asterix.common.exceptions.CompilationException;
-import org.apache.asterix.common.library.ILibraryManager;
import org.apache.asterix.common.transactions.JobId;
import org.apache.asterix.common.utils.StoragePathUtil;
import org.apache.asterix.external.adapter.factory.LookupAdapterFactory;
@@ -127,9 +126,7 @@
import org.apache.hyracks.dataflow.std.file.IFileSplitProvider;
import org.apache.hyracks.dataflow.std.result.ResultWriterOperatorDescriptor;
import org.apache.hyracks.storage.am.btree.dataflow.BTreeSearchOperatorDescriptor;
-import org.apache.hyracks.storage.am.common.api.IMetadataPageManagerFactory;
import org.apache.hyracks.storage.am.common.api.IModificationOperationCallbackFactory;
-import org.apache.hyracks.storage.am.common.api.IPrimitiveValueProviderFactory;
import org.apache.hyracks.storage.am.common.api.ISearchOperationCallbackFactory;
import org.apache.hyracks.storage.am.common.dataflow.IIndexDataflowHelperFactory;
import org.apache.hyracks.storage.am.common.dataflow.IndexDataflowHelperFactory;
@@ -144,10 +141,7 @@
private final ICcApplicationContext appCtx;
private final IStorageComponentProvider storaegComponentProvider;
- private final IMetadataPageManagerFactory metadataPageManagerFactory;
- private final IPrimitiveValueProviderFactory primitiveValueProviderFactory;
private final StorageProperties storageProperties;
- private final ILibraryManager libraryManager;
private final Dataverse defaultDataverse;
private final LockList locks;
@@ -170,9 +164,6 @@
this.defaultDataverse = defaultDataverse;
this.storaegComponentProvider = componentProvider;
storageProperties = appCtx.getStorageProperties();
- libraryManager = appCtx.getLibraryManager();
- metadataPageManagerFactory = componentProvider.getMetadataPageManagerFactory();
- primitiveValueProviderFactory = componentProvider.getPrimitiveValueProviderFactory();
locks = new LockList();
}
@@ -197,10 +188,6 @@
return config;
}
- public ILibraryManager getLibraryManager() {
- return libraryManager;
- }
-
public void setJobId(JobId jobId) {
this.jobId = jobId;
}
@@ -261,10 +248,6 @@
this.resultSerializerFactoryProvider = rafp;
}
- public IResultSerializerFactoryProvider getResultSerializerFactoryProvider() {
- return resultSerializerFactoryProvider;
- }
-
public boolean isWriteTransaction() {
// The transaction writes persistent datasets.
return isWriteTransaction;
@@ -381,10 +364,6 @@
}
}
- public static AlgebricksAbsolutePartitionConstraint determineLocationConstraint(FeedDataSource feedDataSource) {
- return new AlgebricksAbsolutePartitionConstraint(feedDataSource.getLocations());
- }
-
protected Pair<IOperatorDescriptor, AlgebricksPartitionConstraint> buildLoadableDatasetScan(
JobSpecification jobSpec, IAdapterFactory adapterFactory, RecordDescriptor rDesc)
throws AlgebricksException {
diff --git a/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/DatasetUtil.java b/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/DatasetUtil.java
index e2cfb8b..098645e 100644
--- a/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/DatasetUtil.java
+++ b/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/DatasetUtil.java
@@ -53,9 +53,7 @@
import org.apache.hyracks.algebricks.common.exceptions.AlgebricksException;
import org.apache.hyracks.algebricks.common.utils.Pair;
import org.apache.hyracks.algebricks.data.IBinaryComparatorFactoryProvider;
-import org.apache.hyracks.algebricks.data.IBinaryHashFunctionFactoryProvider;
import org.apache.hyracks.api.dataflow.value.IBinaryComparatorFactory;
-import org.apache.hyracks.api.dataflow.value.IBinaryHashFunctionFactory;
import org.apache.hyracks.api.dataflow.value.ISerializerDeserializer;
import org.apache.hyracks.api.dataflow.value.ITypeTraits;
import org.apache.hyracks.api.exceptions.HyracksDataException;
@@ -66,8 +64,8 @@
import org.apache.hyracks.storage.am.common.build.IndexBuilderFactory;
import org.apache.hyracks.storage.am.common.dataflow.IIndexDataflowHelperFactory;
import org.apache.hyracks.storage.am.common.dataflow.IndexCreateOperatorDescriptor;
-import org.apache.hyracks.storage.am.common.dataflow.IndexDropOperatorDescriptor;
import org.apache.hyracks.storage.am.common.dataflow.IndexDataflowHelperFactory;
+import org.apache.hyracks.storage.am.common.dataflow.IndexDropOperatorDescriptor;
import org.apache.hyracks.storage.am.lsm.common.api.ILSMMergePolicyFactory;
import org.apache.hyracks.storage.am.lsm.common.dataflow.LSMTreeIndexCompactOperatorDescriptor;
import org.apache.hyracks.storage.common.IResourceFactory;
@@ -77,99 +75,11 @@
/*
* Dataset related operations
*/
- public static final byte OP_READ = 0x00;
- public static final byte OP_INSERT = 0x01;
- public static final byte OP_DELETE = 0x02;
public static final byte OP_UPSERT = 0x03;
private DatasetUtil() {
}
- public static IBinaryComparatorFactory[] computeKeysBinaryComparatorFactories(Dataset dataset, ARecordType itemType,
- ARecordType metaItemType, IBinaryComparatorFactoryProvider comparatorFactoryProvider)
- throws AlgebricksException {
- List<List<String>> partitioningKeys = dataset.getPrimaryKeys();
- IBinaryComparatorFactory[] bcfs = new IBinaryComparatorFactory[partitioningKeys.size()];
- if (dataset.getDatasetType() == DatasetType.EXTERNAL) {
- // Get comparators for RID fields.
- for (int i = 0; i < partitioningKeys.size(); i++) {
- try {
- bcfs[i] = IndexingConstants.getComparatorFactory(i);
- } catch (AsterixException e) {
- throw new AlgebricksException(e);
- }
- }
- } else {
- InternalDatasetDetails dsd = (InternalDatasetDetails) dataset.getDatasetDetails();
- for (int i = 0; i < partitioningKeys.size(); i++) {
- IAType keyType = (dataset.hasMetaPart() && dsd.getKeySourceIndicator().get(i).intValue() == 1)
- ? metaItemType.getSubFieldType(partitioningKeys.get(i))
- : itemType.getSubFieldType(partitioningKeys.get(i));
- bcfs[i] = comparatorFactoryProvider.getBinaryComparatorFactory(keyType, true);
- }
- }
- return bcfs;
- }
-
- public static int[] createBloomFilterKeyFields(Dataset dataset) throws AlgebricksException {
- if (dataset.getDatasetType() == DatasetType.EXTERNAL) {
- throw new AlgebricksException("not implemented");
- }
- List<List<String>> partitioningKeys = dataset.getPrimaryKeys();
- int[] bloomFilterKeyFields = new int[partitioningKeys.size()];
- for (int i = 0; i < partitioningKeys.size(); ++i) {
- bloomFilterKeyFields[i] = i;
- }
- return bloomFilterKeyFields;
- }
-
- public static IBinaryHashFunctionFactory[] computeKeysBinaryHashFunFactories(Dataset dataset, ARecordType itemType,
- IBinaryHashFunctionFactoryProvider hashFunProvider) throws AlgebricksException {
- if (dataset.getDatasetType() == DatasetType.EXTERNAL) {
- throw new AlgebricksException("not implemented");
- }
- List<List<String>> partitioningKeys = dataset.getPrimaryKeys();
- IBinaryHashFunctionFactory[] bhffs = new IBinaryHashFunctionFactory[partitioningKeys.size()];
- for (int i = 0; i < partitioningKeys.size(); i++) {
- IAType keyType = itemType.getSubFieldType(partitioningKeys.get(i));
- bhffs[i] = hashFunProvider.getBinaryHashFunctionFactory(keyType);
- }
- return bhffs;
- }
-
- public static ITypeTraits[] computeTupleTypeTraits(Dataset dataset, ARecordType itemType, ARecordType metaItemType)
- throws AlgebricksException {
- if (dataset.getDatasetType() == DatasetType.EXTERNAL) {
- throw new AlgebricksException("not implemented");
- }
- List<List<String>> partitioningKeys = dataset.getPrimaryKeys();
- int numKeys = partitioningKeys.size();
- ITypeTraits[] typeTraits;
- if (metaItemType != null) {
- typeTraits = new ITypeTraits[numKeys + 2];
- List<Integer> indicator = ((InternalDatasetDetails) dataset.getDatasetDetails()).getKeySourceIndicator();
- typeTraits[numKeys + 1] = TypeTraitProvider.INSTANCE.getTypeTrait(metaItemType);
- for (int i = 0; i < numKeys; i++) {
- IAType keyType;
- if (indicator.get(i) == 0) {
- keyType = itemType.getSubFieldType(partitioningKeys.get(i));
- } else {
- keyType = metaItemType.getSubFieldType(partitioningKeys.get(i));
- }
- typeTraits[i] = TypeTraitProvider.INSTANCE.getTypeTrait(keyType);
- }
- } else {
- typeTraits = new ITypeTraits[numKeys + 1];
- for (int i = 0; i < numKeys; i++) {
- IAType keyType;
- keyType = itemType.getSubFieldType(partitioningKeys.get(i));
- typeTraits[i] = TypeTraitProvider.INSTANCE.getTypeTrait(keyType);
- }
- }
- typeTraits[numKeys] = TypeTraitProvider.INSTANCE.getTypeTrait(itemType);
- return typeTraits;
- }
-
public static List<String> getFilterField(Dataset dataset) {
return ((InternalDatasetDetails) dataset.getDatasetDetails()).getFilterField();
}
diff --git a/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/IndexUtil.java b/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/IndexUtil.java
index 73964a9..6d07fc7 100644
--- a/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/IndexUtil.java
+++ b/asterixdb/asterix-metadata/src/main/java/org/apache/asterix/metadata/utils/IndexUtil.java
@@ -29,18 +29,11 @@
import org.apache.asterix.metadata.entities.Dataset;
import org.apache.asterix.metadata.entities.Index;
import org.apache.asterix.om.types.ARecordType;
-import org.apache.asterix.runtime.utils.RuntimeUtils;
-import org.apache.hyracks.algebricks.common.constraints.AlgebricksPartitionConstraint;
-import org.apache.hyracks.algebricks.common.constraints.AlgebricksPartitionConstraintHelper;
import org.apache.hyracks.algebricks.common.exceptions.AlgebricksException;
import org.apache.hyracks.algebricks.common.utils.Pair;
import org.apache.hyracks.algebricks.core.rewriter.base.PhysicalOptimizationConfig;
import org.apache.hyracks.api.dataflow.value.ITypeTraits;
import org.apache.hyracks.api.job.JobSpecification;
-import org.apache.hyracks.dataflow.std.file.IFileSplitProvider;
-import org.apache.hyracks.storage.am.common.dataflow.IIndexDataflowHelperFactory;
-import org.apache.hyracks.storage.am.common.dataflow.IndexDropOperatorDescriptor;
-import org.apache.hyracks.storage.am.common.dataflow.IndexDataflowHelperFactory;
public class IndexUtil {
@@ -143,20 +136,6 @@
return secondaryIndexHelper.buildLoadingJobSpec();
}
- public static JobSpecification buildDropSecondaryIndexJobSpec(Index index, MetadataProvider metadataProvider,
- Dataset dataset) throws AlgebricksException {
- JobSpecification spec = RuntimeUtils.createJobSpecification(metadataProvider.getApplicationContext());
- Pair<IFileSplitProvider, AlgebricksPartitionConstraint> splitsAndConstraint =
- metadataProvider.getSplitProviderAndConstraints(dataset, index.getIndexName());
- IIndexDataflowHelperFactory indexHelperFactory = new IndexDataflowHelperFactory(
- metadataProvider.getStorageComponentProvider().getStorageManager(), splitsAndConstraint.first);
- IndexDropOperatorDescriptor btreeDrop = new IndexDropOperatorDescriptor(spec, indexHelperFactory);
- AlgebricksPartitionConstraintHelper.setPartitionConstraintInJobSpec(spec, btreeDrop,
- splitsAndConstraint.second);
- spec.addRoot(btreeDrop);
- return spec;
- }
-
public static JobSpecification buildSecondaryIndexCompactJobSpec(Dataset dataset, Index index, ARecordType recType,
ARecordType metaType, ARecordType enforcedType, ARecordType enforcedMetaType,
MetadataProvider metadataProvider) throws AlgebricksException {