Author: justi9
Date: 2008-09-17 14:45:15 -0400 (Wed, 17 Sep 2008)
New Revision: 2494
Modified:
mgmt/trunk/mint/python/mint/schema.py
mgmt/trunk/mint/sql/schema.sql
Log:
Update to the latest condor schema
Modified: mgmt/trunk/mint/python/mint/schema.py
===================================================================
--- mgmt/trunk/mint/python/mint/schema.py 2008-09-17 18:11:12 UTC (rev 2493)
+++ mgmt/trunk/mint/python/mint/schema.py 2008-09-17 18:45:15 UTC (rev 2494)
@@ -535,6 +535,13 @@
location = StringCol(length=1000, default=None)
defaultInitialFileCount = SmallIntCol(default=None)
defaultDataFileSize = IntCol(default=None)
+ tplIsInitialized = BoolCol(default=None)
+ tplDirectory = StringCol(length=1000, default=None)
+ tplWritePageSize = IntCol(default=None)
+ tplWritePages = IntCol(default=None)
+ tplInitialFileCount = SmallIntCol(default=None)
+ tplDataFileSize = IntCol(default=None)
+ tplCurrentFileCount = IntCol(default=None)
class StoreStats(SQLObject):
@@ -544,6 +551,15 @@
recTime = TimestampCol(default=None)
store = ForeignKey('Store', cascade='null', default=None)
classInfos = dict() # brokerId => classInfo
+ tplTransactionDepth = IntCol(default=None)
+ tplTransactionDepthLow = IntCol(default=None)
+ tplTransactionDepthHigh = IntCol(default=None)
+ tplTxnPrepares = BigIntCol(default=None)
+ tplTxnCommits = BigIntCol(default=None)
+ tplTxnAborts = BigIntCol(default=None)
+ tplOutstandingAIOs = IntCol(default=None)
+ tplOutstandingAIOsLow = IntCol(default=None)
+ tplOutstandingAIOsHigh = IntCol(default=None)
@@ -561,14 +577,17 @@
statsCurr = ForeignKey('JournalStats', cascade='null', default=None)
statsPrev = ForeignKey('JournalStats', cascade='null', default=None)
classInfos = dict() # brokerId => classInfo
+ queue = ForeignKey('Queue', cascade='null', default=None)
name = StringCol(length=1000, default=None)
- queue = ForeignKey('Queue', cascade='null', default=None)
directory = StringCol(length=1000, default=None)
baseFileName = StringCol(length=1000, default=None)
writePageSize = IntCol(default=None)
writePages = IntCol(default=None)
readPageSize = IntCol(default=None)
readPages = IntCol(default=None)
+ initialFileCount = SmallIntCol(default=None)
+ dataFileSize = IntCol(default=None)
+ currentFileCount = IntCol(default=None)
def expand(self, model, callback, by):
@@ -588,14 +607,15 @@
recTime = TimestampCol(default=None)
journal = ForeignKey('Journal', cascade='null', default=None)
classInfos = dict() # brokerId => classInfo
- initialFileCount = SmallIntCol(default=None)
- dataFileSize = IntCol(default=None)
- currentFileCount = IntCol(default=None)
recordDepth = IntCol(default=None)
recordDepthLow = IntCol(default=None)
recordDepthHigh = IntCol(default=None)
- recordEnqueues = BigIntCol(default=None)
- recordDequeues = BigIntCol(default=None)
+ enqueues = BigIntCol(default=None)
+ dequeues = BigIntCol(default=None)
+ txnEnqueues = BigIntCol(default=None)
+ txnDequeues = BigIntCol(default=None)
+ txnCommits = BigIntCol(default=None)
+ txnAborts = BigIntCol(default=None)
outstandingAIOs = IntCol(default=None)
outstandingAIOsLow = IntCol(default=None)
outstandingAIOsHigh = IntCol(default=None)
@@ -780,6 +800,43 @@
conn.callMethod(originalId, classInfo, "GetAd",
callback, args=actualArgs)
+ def SetAttribute(self, model, callback, Name, Value):
+ actualArgs = dict()
+ actualArgs["Name"] = Name
+ actualArgs["Value"] = Value
+ conn = model.connections[self.managedBroker]
+ classInfo = self.classInfos[self.managedBroker]
+ originalId = objectId(None, self.sourceScopeId, self.sourceObjectId)
+ conn.callMethod(originalId, classInfo, "SetAttribute",
+ callback, args=actualArgs)
+
+ def Hold(self, model, callback, Reason):
+ actualArgs = dict()
+ actualArgs["Reason"] = Reason
+ conn = model.connections[self.managedBroker]
+ classInfo = self.classInfos[self.managedBroker]
+ originalId = objectId(None, self.sourceScopeId, self.sourceObjectId)
+ conn.callMethod(originalId, classInfo, "Hold",
+ callback, args=actualArgs)
+
+ def Release(self, model, callback, Reason):
+ actualArgs = dict()
+ actualArgs["Reason"] = Reason
+ conn = model.connections[self.managedBroker]
+ classInfo = self.classInfos[self.managedBroker]
+ originalId = objectId(None, self.sourceScopeId, self.sourceObjectId)
+ conn.callMethod(originalId, classInfo, "Release",
+ callback, args=actualArgs)
+
+ def Remove(self, model, callback, Reason):
+ actualArgs = dict()
+ actualArgs["Reason"] = Reason
+ conn = model.connections[self.managedBroker]
+ classInfo = self.classInfos[self.managedBroker]
+ originalId = objectId(None, self.sourceScopeId, self.sourceObjectId)
+ conn.callMethod(originalId, classInfo, "Remove",
+ callback, args=actualArgs)
+
class JobStats(SQLObject):
class sqlmeta:
lazyUpdate = True
Modified: mgmt/trunk/mint/sql/schema.sql
===================================================================
--- mgmt/trunk/mint/sql/schema.sql 2008-09-17 18:11:12 UTC (rev 2493)
+++ mgmt/trunk/mint/sql/schema.sql 2008-09-17 18:45:15 UTC (rev 2494)
@@ -283,28 +283,32 @@
managed_broker VARCHAR(1000),
stats_curr_id INT,
stats_prev_id INT,
+ queue_id INT,
name VARCHAR(1000),
- queue_id INT,
directory VARCHAR(1000),
base_file_name VARCHAR(1000),
write_page_size INT,
write_pages INT,
read_page_size INT,
- read_pages INT
+ read_pages INT,
+ initial_file_count SMALLINT,
+ data_file_size INT,
+ current_file_count INT
);
CREATE TABLE journal_stats (
id SERIAL PRIMARY KEY,
rec_time TIMESTAMP,
journal_id INT,
- initial_file_count SMALLINT,
- data_file_size INT,
- current_file_count INT,
record_depth INT,
record_depth_low INT,
record_depth_high INT,
- record_enqueues BIGINT,
- record_dequeues BIGINT,
+ enqueues BIGINT,
+ dequeues BIGINT,
+ txn_enqueues BIGINT,
+ txn_dequeues BIGINT,
+ txn_commits BIGINT,
+ txn_aborts BIGINT,
outstanding_ai_os INT,
outstanding_ai_os_low INT,
outstanding_ai_os_high INT,
@@ -593,13 +597,29 @@
broker_id INT,
location VARCHAR(1000),
default_initial_file_count SMALLINT,
- default_data_file_size INT
+ default_data_file_size INT,
+ tpl_is_initialized BOOL,
+ tpl_directory VARCHAR(1000),
+ tpl_write_page_size INT,
+ tpl_write_pages INT,
+ tpl_initial_file_count SMALLINT,
+ tpl_data_file_size INT,
+ tpl_current_file_count INT
);
CREATE TABLE store_stats (
id SERIAL PRIMARY KEY,
rec_time TIMESTAMP,
- store_id INT
+ store_id INT,
+ tpl_transaction_depth INT,
+ tpl_transaction_depth_low INT,
+ tpl_transaction_depth_high INT,
+ tpl_txn_prepares BIGINT,
+ tpl_txn_commits BIGINT,
+ tpl_txn_aborts BIGINT,
+ tpl_outstanding_ai_os INT,
+ tpl_outstanding_ai_os_low INT,
+ tpl_outstanding_ai_os_high INT
);
CREATE TABLE submitter (