57 lines
1.7 KiB
Python
57 lines
1.7 KiB
Python
# vim: tabstop=4 shiftwidth=4 softtabstop=4
|
|
|
|
from keystone.common import sql
|
|
from keystone import exception
|
|
from keystone import token
|
|
|
|
|
|
class TokenModel(sql.ModelBase, sql.DictBase):
|
|
__tablename__ = 'token'
|
|
id = sql.Column(sql.String(64), primary_key=True)
|
|
extra = sql.Column(sql.JsonBlob())
|
|
|
|
@classmethod
|
|
def from_dict(cls, token_dict):
|
|
# shove any non-indexed properties into extra
|
|
data = {}
|
|
token_dict_copy = token_dict.copy()
|
|
data['id'] = token_dict_copy.pop('id')
|
|
data['extra'] = token_dict_copy
|
|
return cls(**data)
|
|
|
|
def to_dict(self):
|
|
extra_copy = self.extra.copy()
|
|
extra_copy['id'] = self.id
|
|
return extra_copy
|
|
|
|
|
|
class Token(sql.Base, token.Driver):
|
|
# Public interface
|
|
def get_token(self, token_id):
|
|
session = self.get_session()
|
|
token_ref = session.query(TokenModel).filter_by(id=token_id).first()
|
|
if not token_ref:
|
|
raise exception.TokenNotFound(token_id=token_id)
|
|
return token_ref.to_dict()
|
|
|
|
def create_token(self, token_id, data):
|
|
data['id'] = token_id
|
|
session = self.get_session()
|
|
with session.begin():
|
|
token_ref = TokenModel.from_dict(data)
|
|
session.add(token_ref)
|
|
session.flush()
|
|
return token_ref.to_dict()
|
|
|
|
def delete_token(self, token_id):
|
|
session = self.get_session()
|
|
token_ref = session.query(TokenModel)\
|
|
.filter_by(id=token_id)\
|
|
.first()
|
|
if not token_ref:
|
|
raise exception.TokenNotFound(token_id=token_id)
|
|
|
|
with session.begin():
|
|
session.delete(token_ref)
|
|
session.flush()
|