X-Git-Url: https://git.jsancho.org/?p=datasette-connectors.git;a=blobdiff_plain;f=datasette_connectors%2Fmonkey.py;fp=datasette_connectors%2Fmonkey.py;h=0fb4e1c7b8d583d4e5e353952bb325122c679c52;hp=e18175f38889f69b8eabc1590f9fc680421b24f9;hb=2bbd8101389969b749f91dea1231682773caebae;hpb=b0802bdb9d86cd65524d6ffa7afb66488d167b1e diff --git a/datasette_connectors/monkey.py b/datasette_connectors/monkey.py index e18175f..0fb4e1c 100644 --- a/datasette_connectors/monkey.py +++ b/datasette_connectors/monkey.py @@ -1,12 +1,15 @@ import asyncio -import datasette -from datasette.app import connections -from datasette.inspect import inspect_hash -from datasette.utils import Results -from pathlib import Path +import threading import sqlite3 -from . import connectors +import datasette.views.base +from datasette.tracer import trace +from datasette.database import Database +from datasette.database import Results + +from .connectors import ConnectorList + +connections = threading.local() def patch_datasette(): @@ -14,74 +17,33 @@ def patch_datasette(): Monkey patching for original Datasette """ - def inspect(self): - " Inspect the database and return a dictionary of table metadata " - if self._inspect: - return self._inspect - - _inspect = {} - files = self.files - - for filename in files: - self.files = (filename,) - path = Path(filename) - name = path.stem - if name in _inspect: - raise Exception("Multiple files with the same stem %s" % name) - try: - _inspect[name] = self.original_inspect()[name] - except sqlite3.DatabaseError: - tables, views, dbtype = connectors.inspect(path) - _inspect[name] = { - "hash": inspect_hash(path), - "file": str(path), - "dbtype": dbtype, - "tables": tables, - "views": views, - } - - self.files = files - self._inspect = _inspect - return self._inspect - - datasette.app.Datasette.original_inspect = datasette.app.Datasette.inspect - datasette.app.Datasette.inspect = inspect + def connect(self, write=False): + try: + # Check if it's a sqlite database + conn = self.original_connect(write=write) + conn.execute("select name from sqlite_master where type='table'") + return conn + except sqlite3.DatabaseError: + conn = ConnectorList.connect(self.path) + return conn + Database.original_connect = Database.connect + Database.connect = connect - async def execute(self, db_name, sql, params=None, truncate=False, custom_time_limit=None, page_size=None): - """Executes sql against db_name in a thread""" - page_size = page_size or self.page_size - def is_sqlite3_conn(): - conn = getattr(connections, db_name, None) + async def execute_fn(self, fn): + def in_thread(): + conn = getattr(connections, self.name, None) if not conn: - info = self.inspect()[db_name] - return info.get('dbtype', 'sqlite3') == 'sqlite3' - else: - return isinstance(conn, sqlite3.Connection) - - def sql_operation_in_thread(): - conn = getattr(connections, db_name, None) - if not conn: - info = self.inspect()[db_name] - conn = connectors.connect(info['file'], info['dbtype']) - setattr(connections, db_name, conn) - - rows, truncated, description = conn.execute( - sql, - params or {}, - truncate=truncate, - page_size=page_size, - max_returned_rows=self.max_returned_rows, - ) - return Results(rows, truncated, description) - - if is_sqlite3_conn(): - return await self.original_execute(db_name, sql, params=params, truncate=truncate, custom_time_limit=custom_time_limit, page_size=page_size) - else: - return await asyncio.get_event_loop().run_in_executor( - self.executor, sql_operation_in_thread - ) - - datasette.app.Datasette.original_execute = datasette.app.Datasette.execute - datasette.app.Datasette.execute = execute + conn = self.connect() + if isinstance(conn, sqlite3.Connection): + self.ds._prepare_connection(conn, self.name) + setattr(connections, self.name, conn) + return fn(conn) + + return await asyncio.get_event_loop().run_in_executor( + self.ds.executor, in_thread + ) + + Database.original_execute_fn = Database.execute_fn + Database.execute_fn = execute_fn