1
0
Fork 0
mirror of https://gitlab.com/bramw/baserow.git synced 2025-04-06 14:05:28 +00:00

Ensure templates and snapshots dont have tsvs made to save disk space.

This commit is contained in:
Nigel Gott 2023-07-06 17:10:51 +01:00
parent 9f6759414c
commit b97d25f4e3
32 changed files with 336 additions and 213 deletions

View file

@ -18,7 +18,7 @@ from baserow.contrib.builder.types import BuilderDict, PageDict
from baserow.contrib.database.constants import IMPORT_SERIALIZED_IMPORTING from baserow.contrib.database.constants import IMPORT_SERIALIZED_IMPORTING
from baserow.core.db import specific_iterator from baserow.core.db import specific_iterator
from baserow.core.models import Application, Workspace from baserow.core.models import Application, Workspace
from baserow.core.registries import ApplicationType, BaserowImportExportMode from baserow.core.registries import ApplicationType, ImportExportConfig
from baserow.core.utils import ChildProgressBuilder from baserow.core.utils import ChildProgressBuilder
@ -86,11 +86,9 @@ class BuilderApplicationType(ApplicationType):
def export_serialized( def export_serialized(
self, self,
builder: Builder, builder: Builder,
import_export_config: ImportExportConfig,
files_zip: Optional[ZipFile] = None, files_zip: Optional[ZipFile] = None,
storage: Optional[Storage] = None, storage: Optional[Storage] = None,
baserow_import_export_mode: Optional[
BaserowImportExportMode
] = BaserowImportExportMode.TARGETING_SAME_WORKSPACE_NEW_PK,
) -> BuilderDict: ) -> BuilderDict:
""" """
Exports the builder application type to a serialized format that can later Exports the builder application type to a serialized format that can later
@ -104,7 +102,7 @@ class BuilderApplicationType(ApplicationType):
serialized_pages = self.export_pages_serialized(pages, files_zip, storage) serialized_pages = self.export_pages_serialized(pages, files_zip, storage)
serialized = super().export_serialized( serialized = super().export_serialized(
builder, files_zip, storage, baserow_import_export_mode builder, import_export_config, files_zip, storage
) )
return BuilderDict(pages=serialized_pages, **serialized) return BuilderDict(pages=serialized_pages, **serialized)
@ -194,6 +192,7 @@ class BuilderApplicationType(ApplicationType):
self, self,
workspace: Workspace, workspace: Workspace,
serialized_values: Dict[str, Any], serialized_values: Dict[str, Any],
import_export_config: ImportExportConfig,
id_mapping: Dict[str, Any], id_mapping: Dict[str, Any],
files_zip: Optional[ZipFile] = None, files_zip: Optional[ZipFile] = None,
storage: Optional[Storage] = None, storage: Optional[Storage] = None,
@ -212,6 +211,7 @@ class BuilderApplicationType(ApplicationType):
application = super().import_serialized( application = super().import_serialized(
workspace, workspace,
serialized_values, serialized_values,
import_export_config,
id_mapping, id_mapping,
files_zip, files_zip,
storage, storage,

View file

@ -12,7 +12,7 @@ from baserow.contrib.builder.domains.models import Domain
from baserow.contrib.builder.exceptions import BuilderDoesNotExist from baserow.contrib.builder.exceptions import BuilderDoesNotExist
from baserow.contrib.builder.models import Builder from baserow.contrib.builder.models import Builder
from baserow.core.exceptions import IdDoesNotExist from baserow.core.exceptions import IdDoesNotExist
from baserow.core.registries import application_type_registry from baserow.core.registries import ImportExportConfig, application_type_registry
from baserow.core.trash.handler import TrashHandler from baserow.core.trash.handler import TrashHandler
from baserow.core.utils import Progress from baserow.core.utils import Progress
@ -163,8 +163,12 @@ class DomainHandler:
builder_application_type = application_type_registry.get("builder") builder_application_type = application_type_registry.get("builder")
import_export_config = ImportExportConfig(
include_permission_data=True, reduce_disk_space_usage=False
)
exported_builder = builder_application_type.export_serialized( exported_builder = builder_application_type.export_serialized(
builder, None, default_storage builder, import_export_config, None, default_storage
) )
progress.increment(by=50) progress.increment(by=50)
@ -173,6 +177,7 @@ class DomainHandler:
duplicate_builder = builder_application_type.import_serialized( duplicate_builder = builder_application_type.import_serialized(
None, None,
exported_builder, exported_builder,
import_export_config,
id_mapping, id_mapping,
None, None,
default_storage, default_storage,

View file

@ -32,6 +32,7 @@ from baserow.contrib.database.views.registries import view_type_registry
from baserow.core.export_serialized import CoreExportSerializedStructure from baserow.core.export_serialized import CoreExportSerializedStructure
from baserow.core.handler import CoreHandler from baserow.core.handler import CoreHandler
from baserow.core.models import Workspace from baserow.core.models import Workspace
from baserow.core.registries import ImportExportConfig
from baserow.core.utils import ChildProgressBuilder, remove_invalid_surrogate_characters from baserow.core.utils import ChildProgressBuilder, remove_invalid_surrogate_characters
from .exceptions import AirtableBaseNotPublic, AirtableShareIsNotABase from .exceptions import AirtableBaseNotPublic, AirtableShareIsNotABase
@ -608,11 +609,18 @@ class AirtableHandler:
download_files_buffer, download_files_buffer,
) )
import_export_config = ImportExportConfig(
# We are not yet downloading any role/permission data from airtable so
# nothing to import
include_permission_data=False,
reduce_disk_space_usage=False,
)
# Import the converted data using the existing method to avoid duplicate code. # Import the converted data using the existing method to avoid duplicate code.
databases, _ = CoreHandler().import_applications_to_workspace( databases, _ = CoreHandler().import_applications_to_workspace(
workspace, workspace,
[baserow_database_export], [baserow_database_export],
files_buffer, files_buffer,
import_export_config,
storage=storage, storage=storage,
progress_builder=progress.create_child_builder(represents_progress=600), progress_builder=progress.create_child_builder(represents_progress=600),
) )

View file

@ -23,7 +23,7 @@ from baserow.contrib.database.views.registries import view_type_registry
from baserow.core.models import Application, Workspace from baserow.core.models import Application, Workspace
from baserow.core.registries import ( from baserow.core.registries import (
ApplicationType, ApplicationType,
BaserowImportExportMode, ImportExportConfig,
serialization_processor_registry, serialization_processor_registry,
) )
from baserow.core.trash.handler import TrashHandler from baserow.core.trash.handler import TrashHandler
@ -70,11 +70,9 @@ class DatabaseApplicationType(ApplicationType):
def export_tables_serialized( def export_tables_serialized(
self, self,
tables: List[Table], tables: List[Table],
import_export_config: ImportExportConfig,
files_zip: Optional[ZipFile] = None, files_zip: Optional[ZipFile] = None,
storage: Optional[Storage] = None, storage: Optional[Storage] = None,
baserow_import_export_mode: Optional[
BaserowImportExportMode
] = BaserowImportExportMode.TARGETING_SAME_WORKSPACE_NEW_PK,
) -> List[Dict[str, Any]]: ) -> List[Dict[str, Any]]:
""" """
Exports the tables provided to a serialized format that can later be Exports the tables provided to a serialized format that can later be
@ -127,26 +125,22 @@ class DatabaseApplicationType(ApplicationType):
views=serialized_views, views=serialized_views,
rows=serialized_rows, rows=serialized_rows,
) )
# Annotate any `SerializationProcessorType` we have.
for ( for serialized_structure in serialization_processor_registry.get_all():
serialized_structure extra_data = serialized_structure.export_serialized(
) in serialization_processor_registry.get_all_for_mode( workspace, table, import_export_config
baserow_import_export_mode
):
structure.update(
**serialized_structure.export_serialized(workspace, table)
) )
if extra_data is not None:
structure.update(**extra_data)
serialized_tables.append(structure) serialized_tables.append(structure)
return serialized_tables return serialized_tables
def export_serialized( def export_serialized(
self, self,
database: Database, database: Database,
import_export_config: ImportExportConfig,
files_zip: Optional[ZipFile] = None, files_zip: Optional[ZipFile] = None,
storage: Optional[Storage] = None, storage: Optional[Storage] = None,
baserow_import_export_mode: Optional[
BaserowImportExportMode
] = BaserowImportExportMode.TARGETING_SAME_WORKSPACE_NEW_PK,
) -> Dict[str, Any]: ) -> Dict[str, Any]:
""" """
Exports the database application type to a serialized format that can later Exports the database application type to a serialized format that can later
@ -161,11 +155,11 @@ class DatabaseApplicationType(ApplicationType):
) )
serialized_tables = self.export_tables_serialized( serialized_tables = self.export_tables_serialized(
tables, files_zip, storage, baserow_import_export_mode tables, import_export_config, files_zip, storage
) )
serialized = super().export_serialized( serialized = super().export_serialized(
database, files_zip, storage, baserow_import_export_mode database, import_export_config, files_zip, storage
) )
serialized.update( serialized.update(
**DatabaseExportSerializedStructure.database(tables=serialized_tables) **DatabaseExportSerializedStructure.database(tables=serialized_tables)
@ -224,13 +218,11 @@ class DatabaseApplicationType(ApplicationType):
database: Database, database: Database,
serialized_tables: List[Dict[str, Any]], serialized_tables: List[Dict[str, Any]],
id_mapping: Dict[str, Any], id_mapping: Dict[str, Any],
import_export_config: ImportExportConfig,
files_zip: Optional[ZipFile] = None, files_zip: Optional[ZipFile] = None,
storage: Optional[Storage] = None, storage: Optional[Storage] = None,
progress_builder: Optional[ChildProgressBuilder] = None, progress_builder: Optional[ChildProgressBuilder] = None,
external_table_fields_to_import: List[Tuple[Table, Dict[str, Any]]] = None, external_table_fields_to_import: List[Tuple[Table, Dict[str, Any]]] = None,
baserow_import_export_mode: Optional[
BaserowImportExportMode
] = BaserowImportExportMode.TARGETING_SAME_WORKSPACE_NEW_PK,
) -> List[Table]: ) -> List[Table]:
""" """
Imports tables exported by the `export_tables_serialized` method. Look at Imports tables exported by the `export_tables_serialized` method. Look at
@ -252,9 +244,8 @@ class DatabaseApplicationType(ApplicationType):
field to import. field to import.
Useful for when importing a single table which also needs to add related Useful for when importing a single table which also needs to add related
fields to other existing tables in the database. fields to other existing tables in the database.
:param baserow_import_export_mode: defines which Baserow import/export mode to :param import_export_config: provides configuration options for the
use, defaults to `TARGETING_SAME_WORKSPACE_NEW_PK`. import/export process to customize how it works.
:type baserow_import_export_mode: enum
:return: The list of created tables :return: The list of created tables
""" """
@ -302,6 +293,7 @@ class DatabaseApplicationType(ApplicationType):
field_instance = field_type.import_serialized( field_instance = field_type.import_serialized(
serialized_table["_object"], serialized_table["_object"],
serialized_field, serialized_field,
import_export_config,
id_mapping, id_mapping,
deferred_fk_update_collector, deferred_fk_update_collector,
) )
@ -315,6 +307,7 @@ class DatabaseApplicationType(ApplicationType):
external_field = field_type.import_serialized( external_field = field_type.import_serialized(
external_table, external_table,
serialized_field, serialized_field,
import_export_config,
id_mapping, id_mapping,
deferred_fk_update_collector, deferred_fk_update_collector,
) )
@ -494,14 +487,13 @@ class DatabaseApplicationType(ApplicationType):
source_workspace = Workspace.objects.get(pk=id_mapping["import_workspace_id"]) source_workspace = Workspace.objects.get(pk=id_mapping["import_workspace_id"])
for serialized_table in serialized_tables: for serialized_table in serialized_tables:
table = serialized_table["_object"] table = serialized_table["_object"]
SearchHandler.entire_field_values_changed_or_created(table) if not import_export_config.reduce_disk_space_usage:
SearchHandler.entire_field_values_changed_or_created(table)
for ( for (
serialized_structure serialized_structure_processor
) in serialization_processor_registry.get_all_for_mode( ) in serialization_processor_registry.get_all():
baserow_import_export_mode serialized_structure_processor.import_serialized(
): source_workspace, table, serialized_table, import_export_config
serialized_structure.import_serialized(
source_workspace, table, serialized_table
) )
return imported_tables return imported_tables
@ -510,13 +502,11 @@ class DatabaseApplicationType(ApplicationType):
self, self,
workspace: Workspace, workspace: Workspace,
serialized_values: Dict[str, Any], serialized_values: Dict[str, Any],
import_export_config: ImportExportConfig,
id_mapping: Dict[str, Any], id_mapping: Dict[str, Any],
files_zip: Optional[ZipFile] = None, files_zip: Optional[ZipFile] = None,
storage: Optional[Storage] = None, storage: Optional[Storage] = None,
progress_builder: Optional[ChildProgressBuilder] = None, progress_builder: Optional[ChildProgressBuilder] = None,
baserow_import_export_mode: Optional[
BaserowImportExportMode
] = BaserowImportExportMode.TARGETING_SAME_WORKSPACE_NEW_PK,
) -> Application: ) -> Application:
""" """
Imports a database application exported by the `export_serialized` method. Imports a database application exported by the `export_serialized` method.
@ -530,11 +520,11 @@ class DatabaseApplicationType(ApplicationType):
application = super().import_serialized( application = super().import_serialized(
workspace, workspace,
serialized_values, serialized_values,
import_export_config,
id_mapping, id_mapping,
files_zip, files_zip,
storage, storage,
progress.create_child_builder(represents_progress=database_progress), progress.create_child_builder(represents_progress=database_progress),
baserow_import_export_mode=baserow_import_export_mode,
) )
database = application.specific database = application.specific
@ -546,10 +536,10 @@ class DatabaseApplicationType(ApplicationType):
database, database,
serialized_values["tables"], serialized_values["tables"],
id_mapping, id_mapping,
import_export_config,
files_zip, files_zip,
storage, storage,
progress.create_child_builder(represents_progress=table_progress), progress.create_child_builder(represents_progress=table_progress),
baserow_import_export_mode=baserow_import_export_mode,
) )
return database return database

View file

@ -83,12 +83,14 @@ from baserow.contrib.database.validators import UnicodeRegexValidator
from baserow.core.fields import SyncedDateTimeField from baserow.core.fields import SyncedDateTimeField
from baserow.core.handler import CoreHandler from baserow.core.handler import CoreHandler
from baserow.core.models import UserFile, WorkspaceUser from baserow.core.models import UserFile, WorkspaceUser
from baserow.core.registries import ImportExportConfig
from baserow.core.user_files.exceptions import UserFileDoesNotExist from baserow.core.user_files.exceptions import UserFileDoesNotExist
from baserow.core.user_files.handler import UserFileHandler from baserow.core.user_files.handler import UserFileHandler
from baserow.core.utils import list_to_comma_separated_string from baserow.core.utils import list_to_comma_separated_string
from baserow.formula import BaserowFormulaException from baserow.formula import BaserowFormulaException
from baserow.formula.exceptions import FormulaFunctionTypeDoesNotExist from baserow.formula.exceptions import FormulaFunctionTypeDoesNotExist
from ..search.handler import SearchHandler
from .constants import UPSERT_OPTION_DICT_KEY from .constants import UPSERT_OPTION_DICT_KEY
from .deferred_field_fk_updater import DeferredFieldFkUpdater from .deferred_field_fk_updater import DeferredFieldFkUpdater
from .dependencies.exceptions import ( from .dependencies.exceptions import (
@ -2002,6 +2004,7 @@ class LinkRowFieldType(FieldType):
self, self,
table: "Table", table: "Table",
serialized_values: Dict[str, Any], serialized_values: Dict[str, Any],
import_export_config: ImportExportConfig,
id_mapping: Dict[str, Any], id_mapping: Dict[str, Any],
deferred_fk_update_collector: DeferredFieldFkUpdater, deferred_fk_update_collector: DeferredFieldFkUpdater,
) -> Optional[Field]: ) -> Optional[Field]:
@ -2034,7 +2037,11 @@ class LinkRowFieldType(FieldType):
serialized_copy["link_row_relation_id"] = related_field.link_row_relation_id serialized_copy["link_row_relation_id"] = related_field.link_row_relation_id
field = super().import_serialized( field = super().import_serialized(
table, serialized_copy, id_mapping, deferred_fk_update_collector table,
serialized_copy,
import_export_config,
id_mapping,
deferred_fk_update_collector,
) )
if related_field_found: if related_field_found:
@ -3498,6 +3505,7 @@ class FormulaFieldType(ReadOnlyFieldType):
if should_send_signals_at_end: if should_send_signals_at_end:
update_collector.apply_updates_and_get_updated_fields(field_cache) update_collector.apply_updates_and_get_updated_fields(field_cache)
SearchHandler().entire_field_values_changed_or_created(field.table, [field])
update_collector.send_force_refresh_signals_for_all_updated_tables() update_collector.send_force_refresh_signals_for_all_updated_tables()
def row_of_dependency_updated( def row_of_dependency_updated(
@ -3810,6 +3818,7 @@ class CountFieldType(FormulaFieldType):
self, self,
table: "Table", table: "Table",
serialized_values: Dict[str, Any], serialized_values: Dict[str, Any],
import_export_config: ImportExportConfig,
id_mapping: Dict[str, Any], id_mapping: Dict[str, Any],
deferred_fk_update_collector: DeferredFieldFkUpdater, deferred_fk_update_collector: DeferredFieldFkUpdater,
) -> "Field": ) -> "Field":
@ -3819,7 +3828,11 @@ class CountFieldType(FormulaFieldType):
# the mapping. # the mapping.
original_through_field_id = serialized_copy.pop("through_field_id") original_through_field_id = serialized_copy.pop("through_field_id")
field = super().import_serialized( field = super().import_serialized(
table, serialized_copy, id_mapping, deferred_fk_update_collector table,
serialized_copy,
import_export_config,
id_mapping,
deferred_fk_update_collector,
) )
deferred_fk_update_collector.add_deferred_fk_to_update( deferred_fk_update_collector.add_deferred_fk_to_update(
field, "through_field_id", original_through_field_id field, "through_field_id", original_through_field_id
@ -3956,6 +3969,7 @@ class RollupFieldType(FormulaFieldType):
self, self,
table: "Table", table: "Table",
serialized_values: Dict[str, Any], serialized_values: Dict[str, Any],
import_export_config: ImportExportConfig,
id_mapping: Dict[str, Any], id_mapping: Dict[str, Any],
deferred_fk_update_collector: DeferredFieldFkUpdater, deferred_fk_update_collector: DeferredFieldFkUpdater,
) -> "Field": ) -> "Field":
@ -3966,7 +3980,11 @@ class RollupFieldType(FormulaFieldType):
original_through_field_id = serialized_copy.pop("through_field_id") original_through_field_id = serialized_copy.pop("through_field_id")
original_target_field_id = serialized_copy.pop("target_field_id") original_target_field_id = serialized_copy.pop("target_field_id")
field = super().import_serialized( field = super().import_serialized(
table, serialized_copy, id_mapping, deferred_fk_update_collector table,
serialized_copy,
import_export_config,
id_mapping,
deferred_fk_update_collector,
) )
deferred_fk_update_collector.add_deferred_fk_to_update( deferred_fk_update_collector.add_deferred_fk_to_update(
field, "through_field_id", original_through_field_id field, "through_field_id", original_through_field_id
@ -4192,6 +4210,7 @@ class LookupFieldType(FormulaFieldType):
self, self,
table: "Table", table: "Table",
serialized_values: Dict[str, Any], serialized_values: Dict[str, Any],
import_export_config: ImportExportConfig,
id_mapping: Dict[str, Any], id_mapping: Dict[str, Any],
deferred_fk_update_collector: DeferredFieldFkUpdater, deferred_fk_update_collector: DeferredFieldFkUpdater,
) -> "Field": ) -> "Field":
@ -4202,7 +4221,11 @@ class LookupFieldType(FormulaFieldType):
original_through_field_id = serialized_copy.pop("through_field_id") original_through_field_id = serialized_copy.pop("through_field_id")
original_target_field_id = serialized_copy.pop("target_field_id") original_target_field_id = serialized_copy.pop("target_field_id")
field = super().import_serialized( field = super().import_serialized(
table, serialized_copy, id_mapping, deferred_fk_update_collector table,
serialized_copy,
import_export_config,
id_mapping,
deferred_fk_update_collector,
) )
deferred_fk_update_collector.add_deferred_fk_to_update( deferred_fk_update_collector.add_deferred_fk_to_update(
field, "through_field_id", original_through_field_id field, "through_field_id", original_through_field_id

View file

@ -18,6 +18,7 @@ from django.db.models.functions import Cast
from baserow.contrib.database.fields.constants import UPSERT_OPTION_DICT_KEY from baserow.contrib.database.fields.constants import UPSERT_OPTION_DICT_KEY
from baserow.contrib.database.fields.field_sortings import OptionallyAnnotatedOrderBy from baserow.contrib.database.fields.field_sortings import OptionallyAnnotatedOrderBy
from baserow.core.registries import ImportExportConfig
from baserow.core.registry import ( from baserow.core.registry import (
APIUrlsInstanceMixin, APIUrlsInstanceMixin,
APIUrlsRegistryMixin, APIUrlsRegistryMixin,
@ -786,6 +787,7 @@ class FieldType(
self, self,
table: "Table", table: "Table",
serialized_values: Dict[str, Any], serialized_values: Dict[str, Any],
import_export_config: ImportExportConfig,
id_mapping: Dict[str, Any], id_mapping: Dict[str, Any],
deferred_fk_update_collector: DeferredFieldFkUpdater, deferred_fk_update_collector: DeferredFieldFkUpdater,
) -> Field: ) -> Field:
@ -798,6 +800,8 @@ class FieldType(
be imported. be imported.
:param id_mapping: The map of exported ids to newly created ids that must be :param id_mapping: The map of exported ids to newly created ids that must be
updated when a new instance has been created. updated when a new instance has been created.
:param import_export_config: provides configuration options for the
import/export process to customize how it works.
:param deferred_fk_update_collector: An object than can be used to defer :param deferred_fk_update_collector: An object than can be used to defer
setting FK's to other fields until after all fields have been created setting FK's to other fields until after all fields have been created
and we know their IDs. and we know their IDs.
@ -818,9 +822,10 @@ class FieldType(
if self.can_have_select_options if self.can_have_select_options
else [] else []
) )
should_create_tsvector_column = not import_export_config.reduce_disk_space_usage
field = self.model_class( field = self.model_class(
table=table, table=table,
tsvector_column_created=table.tsvectors_are_supported, tsvector_column_created=should_create_tsvector_column,
**serialized_copy, **serialized_copy,
) )
field.save() field.save()

View file

@ -338,7 +338,7 @@ class SearchHandler(
and field_ids_to_restrict_update_to is not None and field_ids_to_restrict_update_to is not None
): ):
raise ValueError( raise ValueError(
"Mst always update all fields when updating rows " "Must always update all fields when updating rows "
"with needs_background_update=True." "with needs_background_update=True."
) )
@ -369,6 +369,7 @@ class SearchHandler(
was_full_column_update = not update_tsvectors_for_changed_rows_only was_full_column_update = not update_tsvectors_for_changed_rows_only
if ( if (
was_full_column_update was_full_column_update
and collected_vectors
and settings.AUTO_VACUUM_AFTER_SEARCH_UPDATE and settings.AUTO_VACUUM_AFTER_SEARCH_UPDATE
and not settings.TESTS and not settings.TESTS
): ):

View file

@ -32,7 +32,7 @@ from baserow.contrib.database.rows.handler import RowHandler
from baserow.contrib.database.views.handler import ViewHandler from baserow.contrib.database.views.handler import ViewHandler
from baserow.contrib.database.views.view_types import GridViewType from baserow.contrib.database.views.view_types import GridViewType
from baserow.core.handler import CoreHandler from baserow.core.handler import CoreHandler
from baserow.core.registries import application_type_registry from baserow.core.registries import ImportExportConfig, application_type_registry
from baserow.core.telemetry.utils import baserow_trace_methods from baserow.core.telemetry.utils import baserow_trace_methods
from baserow.core.trash.handler import TrashHandler from baserow.core.trash.handler import TrashHandler
from baserow.core.utils import ChildProgressBuilder, Progress, find_unused_name from baserow.core.utils import ChildProgressBuilder, Progress, find_unused_name
@ -533,7 +533,11 @@ class TableHandler(metaclass=baserow_trace_methods(tracer)):
database_type = application_type_registry.get_by_model(database) database_type = application_type_registry.get_by_model(database)
serialized_tables = database_type.export_tables_serialized([table]) config = ImportExportConfig(
include_permission_data=True, reduce_disk_space_usage=False
)
serialized_tables = database_type.export_tables_serialized([table], config)
# Set a unique name for the table to import back as a new one. # Set a unique name for the table to import back as a new one.
exported_table = serialized_tables[0] exported_table = serialized_tables[0]
@ -570,6 +574,7 @@ class TableHandler(metaclass=baserow_trace_methods(tracer)):
database, database,
[exported_table], [exported_table],
id_mapping, id_mapping,
config,
external_table_fields_to_import=link_fields_to_import_to_existing_tables, external_table_fields_to_import=link_fields_to_import_to_existing_tables,
progress_builder=progress.create_child_builder( progress_builder=progress.create_child_builder(
represents_progress=import_progress represents_progress=import_progress

View file

@ -72,7 +72,7 @@ from .operations import (
UpdateWorkspaceUserOperationType, UpdateWorkspaceUserOperationType,
) )
from .registries import ( from .registries import (
BaserowImportExportMode, ImportExportConfig,
application_type_registry, application_type_registry,
object_scope_type_registry, object_scope_type_registry,
operation_type_registry, operation_type_registry,
@ -1355,11 +1355,16 @@ class CoreHandler(metaclass=baserow_trace_methods(tracer)):
progress = ChildProgressBuilder.build(progress_builder, child_total=100) progress = ChildProgressBuilder.build(progress_builder, child_total=100)
progress.increment(by=start_progress) progress.increment(by=start_progress)
duplicate_import_export_config = ImportExportConfig(
include_permission_data=True, reduce_disk_space_usage=False
)
# export the application # export the application
specific_application = application.specific specific_application = application.specific
application_type = application_type_registry.get_by_model(specific_application) application_type = application_type_registry.get_by_model(specific_application)
try: try:
serialized = application_type.export_serialized(specific_application) serialized = application_type.export_serialized(
specific_application, duplicate_import_export_config
)
except OperationalError as e: except OperationalError as e:
# Detect if this `OperationalError` is due to us exceeding the # Detect if this `OperationalError` is due to us exceeding the
# lock count in `max_locks_per_transaction`. If it is, we'll # lock count in `max_locks_per_transaction`. If it is, we'll
@ -1382,6 +1387,7 @@ class CoreHandler(metaclass=baserow_trace_methods(tracer)):
new_application_clone = application_type.import_serialized( new_application_clone = application_type.import_serialized(
workspace, workspace,
serialized, serialized,
duplicate_import_export_config,
id_mapping, id_mapping,
progress_builder=progress.create_child_builder( progress_builder=progress.create_child_builder(
represents_progress=import_progress represents_progress=import_progress
@ -1478,10 +1484,8 @@ class CoreHandler(metaclass=baserow_trace_methods(tracer)):
self, self,
workspace, workspace,
files_buffer, files_buffer,
import_export_config: ImportExportConfig,
storage=None, storage=None,
baserow_import_export_mode: Optional[
BaserowImportExportMode
] = BaserowImportExportMode.TARGETING_DIFF_WORKSPACE_NEW_PK,
): ):
""" """
Exports the applications of a workspace to a list. They can later be imported Exports the applications of a workspace to a list. They can later be imported
@ -1498,9 +1502,8 @@ class CoreHandler(metaclass=baserow_trace_methods(tracer)):
:type files_buffer: IOBase :type files_buffer: IOBase
:param storage: The storage where the files can be loaded from. :param storage: The storage where the files can be loaded from.
:type storage: Storage or None :type storage: Storage or None
:param baserow_import_export_mode: defines which Baserow import/export mode to :param import_export_config: provides configuration options for the
use, defaults to `TARGETING_DIFF_WORKSPACE_NEW_PK`. import/export process to customize how it works.
:type baserow_import_export_mode: enum
:return: A list containing the exported applications. :return: A list containing the exported applications.
:rtype: list :rtype: list
""" """
@ -1516,7 +1519,7 @@ class CoreHandler(metaclass=baserow_trace_methods(tracer)):
application_type = application_type_registry.get_by_model(application) application_type = application_type_registry.get_by_model(application)
with application_type.export_safe_transaction_context(application): with application_type.export_safe_transaction_context(application):
exported_application = application_type.export_serialized( exported_application = application_type.export_serialized(
application, files_zip, storage, baserow_import_export_mode application, import_export_config, files_zip, storage
) )
exported_applications.append(exported_application) exported_applications.append(exported_application)
@ -1527,11 +1530,9 @@ class CoreHandler(metaclass=baserow_trace_methods(tracer)):
workspace: Workspace, workspace: Workspace,
exported_applications: List[Dict[str, Any]], exported_applications: List[Dict[str, Any]],
files_buffer: IO[bytes], files_buffer: IO[bytes],
import_export_config: ImportExportConfig,
storage: Optional[Storage] = None, storage: Optional[Storage] = None,
progress_builder: Optional[ChildProgressBuilder] = None, progress_builder: Optional[ChildProgressBuilder] = None,
baserow_import_export_mode: Optional[
BaserowImportExportMode
] = BaserowImportExportMode.TARGETING_DIFF_WORKSPACE_NEW_PK,
) -> Tuple[List[Application], Dict[str, Any]]: ) -> Tuple[List[Application], Dict[str, Any]]:
""" """
Imports multiple exported applications into the given workspace. It is Imports multiple exported applications into the given workspace. It is
@ -1547,9 +1548,8 @@ class CoreHandler(metaclass=baserow_trace_methods(tracer)):
:param storage: The storage where the files can be copied to. :param storage: The storage where the files can be copied to.
:param progress_builder: If provided will be used to build a child progress bar :param progress_builder: If provided will be used to build a child progress bar
and report on this methods progress to the parent of the progress_builder. and report on this methods progress to the parent of the progress_builder.
:param baserow_import_export_mode: defines which Baserow import/export mode to :param import_export_config: provides configuration options for the
use, defaults to `TARGETING_DIFF_WORKSPACE_NEW_PK`. import/export process to customize how it works.
:type baserow_import_export_mode: enum
:return: The newly created applications based on the import and a dict :return: The newly created applications based on the import and a dict
containing a mapping of old ids to new ids. containing a mapping of old ids to new ids.
""" """
@ -1570,13 +1570,13 @@ class CoreHandler(metaclass=baserow_trace_methods(tracer)):
imported_application = application_type.import_serialized( imported_application = application_type.import_serialized(
workspace, workspace,
application, application,
import_export_config,
id_mapping, id_mapping,
files_zip, files_zip,
storage, storage,
progress_builder=progress.create_child_builder( progress_builder=progress.create_child_builder(
represents_progress=1000 represents_progress=1000
), ),
baserow_import_export_mode=baserow_import_export_mode,
) )
imported_application.order = next_application_order_value imported_application.order = next_application_order_value
next_application_order_value += 1 next_application_order_value += 1
@ -1644,6 +1644,13 @@ class CoreHandler(metaclass=baserow_trace_methods(tracer)):
) )
installed_categories = list(TemplateCategory.objects.all()) installed_categories = list(TemplateCategory.objects.all())
sync_templates_import_export_config = ImportExportConfig(
include_permission_data=False,
# Without reducing disk space usage Baserow after first time install
# takes up over 1GB of disk space.
reduce_disk_space_usage=True,
)
# Loop over the JSON template files in the directory to see which database # Loop over the JSON template files in the directory to see which database
# templates need to be created or updated. # templates need to be created or updated.
templates = list( templates = list(
@ -1701,6 +1708,7 @@ class CoreHandler(metaclass=baserow_trace_methods(tracer)):
workspace, workspace,
parsed_json["export"], parsed_json["export"],
files_buffer=files_buffer, files_buffer=files_buffer,
import_export_config=sync_templates_import_export_config,
storage=storage, storage=storage,
) )
@ -1820,6 +1828,9 @@ class CoreHandler(metaclass=baserow_trace_methods(tracer)):
workspace, workspace,
parsed_json["export"], parsed_json["export"],
files_buffer=files_buffer, files_buffer=files_buffer,
import_export_config=ImportExportConfig(
include_permission_data=False, reduce_disk_space_usage=False
),
storage=storage, storage=storage,
progress_builder=progress_builder, progress_builder=progress_builder,
) )

View file

@ -6,6 +6,11 @@ from django.core.management.base import BaseCommand
from baserow.core.handler import CoreHandler from baserow.core.handler import CoreHandler
from baserow.core.models import Workspace from baserow.core.models import Workspace
from baserow.core.registries import ImportExportConfig
cli_import_export_config = ImportExportConfig(
include_permission_data=False, reduce_disk_space_usage=False
)
class Command(BaseCommand): class Command(BaseCommand):
@ -62,6 +67,7 @@ class Command(BaseCommand):
exported_applications = CoreHandler().export_workspace_applications( exported_applications = CoreHandler().export_workspace_applications(
workspace, workspace,
files_buffer=files_buffer, files_buffer=files_buffer,
import_export_config=cli_import_export_config,
) )
with open(export_path, "w") as export_buffer: with open(export_path, "w") as export_buffer:

View file

@ -6,6 +6,9 @@ from django.core.management.base import BaseCommand
from django.db import transaction from django.db import transaction
from baserow.core.handler import CoreHandler from baserow.core.handler import CoreHandler
from baserow.core.management.commands.export_workspace_applications import (
cli_import_export_config,
)
from baserow.core.models import Workspace from baserow.core.models import Workspace
@ -66,9 +69,7 @@ class Command(BaseCommand):
# By default, we won't import any registry data. This is because # By default, we won't import any registry data. This is because
# `RoleAssignment` can't be imported if the subjects are teams. # `RoleAssignment` can't be imported if the subjects are teams.
applications, _ = handler.import_applications_to_workspace( applications, _ = handler.import_applications_to_workspace(
workspace, workspace, content, files_buffer, cli_import_export_config
content,
files_buffer,
) )
if files_buffer: if files_buffer:

View file

@ -1,6 +1,6 @@
import abc import abc
import dataclasses
from collections import defaultdict from collections import defaultdict
from enum import Enum
from functools import cached_property from functools import cached_property
from typing import TYPE_CHECKING, Any, Dict, Iterable, List, Optional, TypeVar, Union from typing import TYPE_CHECKING, Any, Dict, Iterable, List, Optional, TypeVar, Union
from xmlrpc.client import Boolean from xmlrpc.client import Boolean
@ -33,9 +33,7 @@ from .export_serialized import CoreExportSerializedStructure
from .registry import ( from .registry import (
APIUrlsInstanceMixin, APIUrlsInstanceMixin,
APIUrlsRegistryMixin, APIUrlsRegistryMixin,
ImportExportMixin,
Instance, Instance,
InstanceSubClass,
MapAPIExceptionsInstanceMixin, MapAPIExceptionsInstanceMixin,
ModelInstanceMixin, ModelInstanceMixin,
ModelRegistryMixin, ModelRegistryMixin,
@ -61,19 +59,29 @@ if TYPE_CHECKING:
) )
class BaserowImportExportMode(Enum): @dataclasses.dataclass
""" class ImportExportConfig:
Responsible for informing our import and export functionality
what kind of mode to support:
targetingSameWorkspaceWithNewIds: we are exporting and importing into """
the same workspace, with new primary keys (not preserving IDs). When true the export/import will also transfer any permission data.
targetingDifferentWorkspaceWithNewIds: we are exporting and importing
into a new workspace, with new primary keys (not preserving IDs). For example when exporting to JSON we don't want to include RBAC data as we would
also need to export all the subjects, so setting this to False will exclude
RBAC roles from the export.
""" """
TARGETING_SAME_WORKSPACE_NEW_PK = "targetingSameWorkspaceWithNewIds" include_permission_data: bool
TARGETING_DIFF_WORKSPACE_NEW_PK = "targetingDifferentWorkspaceWithNewIds"
"""
Whether or not the import/export should attempt to save disk space by excluding
certain pieces of optional data or processes that could instead be done later or
not used at all.
For example, this configures the database when True to not create/populate
tsvector full text search columns as they can also be lazy loaded after the import
when the user opens a view.
"""
reduce_disk_space_usage: bool = False
class Plugin(APIUrlsInstanceMixin, Instance): class Plugin(APIUrlsInstanceMixin, Instance):
@ -199,7 +207,6 @@ class PluginRegistry(APIUrlsRegistryMixin, Registry):
class ApplicationType( class ApplicationType(
APIUrlsInstanceMixin, APIUrlsInstanceMixin,
ModelInstanceMixin["Application"], ModelInstanceMixin["Application"],
ImportExportMixin["Application"],
Instance, Instance,
): ):
""" """
@ -302,7 +309,7 @@ class ApplicationType(
workspace: "Workspace", workspace: "Workspace",
scope, scope,
exported_structure: dict, exported_structure: dict,
baserow_import_export_mode: BaserowImportExportMode, import_export_config: ImportExportConfig,
) -> dict: ) -> dict:
""" """
Given a serialized dictionary generated by `export_serialized`, this method Given a serialized dictionary generated by `export_serialized`, this method
@ -310,12 +317,12 @@ class ApplicationType(
that needs to be added to the serialized structure. that needs to be added to the serialized structure.
""" """
for serialized_structure in serialization_processor_registry.get_all_for_mode( for serialized_structure in serialization_processor_registry.get_all():
baserow_import_export_mode data = serialized_structure.export_serialized(
): workspace, scope, import_export_config
exported_structure.update(
**serialized_structure.export_serialized(workspace, scope)
) )
if data is not None:
exported_structure.update(**data)
return exported_structure return exported_structure
def import_serialized_structure_with_registry( def import_serialized_structure_with_registry(
@ -323,7 +330,7 @@ class ApplicationType(
id_mapping: Dict[str, Any], id_mapping: Dict[str, Any],
scope, scope,
serialized_scope: dict, serialized_scope: dict,
baserow_import_export_mode: BaserowImportExportMode, import_export_config: ImportExportConfig,
workspace: Optional["Workspace"] = None, workspace: Optional["Workspace"] = None,
) -> None: ) -> None:
""" """
@ -340,21 +347,17 @@ class ApplicationType(
pk=id_mapping["import_workspace_id"] pk=id_mapping["import_workspace_id"]
) )
for serialized_structure in serialization_processor_registry.get_all_for_mode( for serialized_structure in serialization_processor_registry.get_all():
baserow_import_export_mode
):
serialized_structure.import_serialized( serialized_structure.import_serialized(
source_workspace, scope, serialized_scope source_workspace, scope, serialized_scope, import_export_config
) )
def export_serialized( def export_serialized(
self, self,
application: "Application", application: "Application",
import_export_config: ImportExportConfig,
files_zip: Optional[ZipFile] = None, files_zip: Optional[ZipFile] = None,
storage: Optional[Storage] = None, storage: Optional[Storage] = None,
baserow_import_export_mode: Optional[
BaserowImportExportMode
] = BaserowImportExportMode.TARGETING_SAME_WORKSPACE_NEW_PK,
): ):
""" """
Exports the application to a serialized dict that can be imported by the Exports the application to a serialized dict that can be imported by the
@ -367,9 +370,8 @@ class ApplicationType(
:type files_zip: ZipFile :type files_zip: ZipFile
:param storage: The storage where the files can be loaded from. :param storage: The storage where the files can be loaded from.
:type storage: Storage or None :type storage: Storage or None
:param baserow_import_export_mode: defines which Baserow import/export mode to :param import_export_config: provides configuration options for the
use, defaults to `TARGETING_SAME_WORKSPACE_NEW_PK`. import/export process to customize how it works.
:type baserow_import_export_mode: enum
:return: The exported and serialized application. :return: The exported and serialized application.
:rtype: dict :rtype: dict
""" """
@ -382,7 +384,7 @@ class ApplicationType(
) )
# Annotate any `SerializationProcessorType` we have. # Annotate any `SerializationProcessorType` we have.
structure = self.export_serialized_structure_with_registry( structure = self.export_serialized_structure_with_registry(
application.get_root(), application, structure, baserow_import_export_mode application.get_root(), application, structure, import_export_config
) )
return structure return structure
@ -390,13 +392,11 @@ class ApplicationType(
self, self,
workspace: "Workspace", workspace: "Workspace",
serialized_values: Dict[str, Any], serialized_values: Dict[str, Any],
import_export_config: ImportExportConfig,
id_mapping: Dict[str, Any], id_mapping: Dict[str, Any],
files_zip: Optional[ZipFile] = None, files_zip: Optional[ZipFile] = None,
storage: Optional[Storage] = None, storage: Optional[Storage] = None,
progress_builder: Optional[ChildProgressBuilder] = None, progress_builder: Optional[ChildProgressBuilder] = None,
baserow_import_export_mode: Optional[
BaserowImportExportMode
] = BaserowImportExportMode.TARGETING_SAME_WORKSPACE_NEW_PK,
) -> "Application": ) -> "Application":
""" """
Imports the exported serialized application by the `export_serialized` as a new Imports the exported serialized application by the `export_serialized` as a new
@ -412,9 +412,8 @@ class ApplicationType(
:param storage: The storage where the files can be copied to. :param storage: The storage where the files can be copied to.
:param progress_builder: If provided will be used to build a child progress bar :param progress_builder: If provided will be used to build a child progress bar
and report on this methods progress to the parent of the progress_builder. and report on this methods progress to the parent of the progress_builder.
:param baserow_import_export_mode: defines which Baserow import/export mode to :param import_export_config: provides configuration options for the
use, defaults to `TARGETING_SAME_WORKSPACE_NEW_PK`. import/export process to customize how it works.
:type baserow_import_export_mode: enum
:return: The newly created application. :return: The newly created application.
""" """
@ -460,7 +459,7 @@ class ApplicationType(
id_mapping, id_mapping,
application, application,
serialized_values, serialized_values,
baserow_import_export_mode, import_export_config,
workspace, workspace,
) )
@ -1158,34 +1157,34 @@ class SerializationProcessorType(abc.ABC, Instance):
`import_serialized` and `export_serialized` methods. `import_serialized` and `export_serialized` methods.
""" """
import_export_mode = None
@classmethod @classmethod
def import_serialized( def import_serialized(
cls, cls,
workspace: "Workspace", workspace: "Workspace",
scope: SerializationProcessorScope, scope: SerializationProcessorScope,
serialized_scope: dict, serialized_scope: dict,
) -> dict: import_export_config: ImportExportConfig,
):
""" """
A hook which is called after an application subclass or table has been A hook which is called after an application subclass or table has been
imported, allowing us to import additional data in `serialized_scope`. imported, allowing us to import additional data in `serialized_scope`.
""" """
return {} pass
@classmethod @classmethod
def export_serialized( def export_serialized(
cls, cls,
workspace: "Workspace", workspace: "Workspace",
scope: SerializationProcessorScope, scope: SerializationProcessorScope,
) -> dict[str, Any]: import_export_config: ImportExportConfig,
) -> Optional[Dict[str, Any]]:
""" """
A hook which is called after an application subclass or table has been A hook which is called after an application subclass or table has been
exported, allowing us to export additional data. exported, allowing us to export additional data.
""" """
return {} return None
class SerializationProcessorRegistry(Registry[SerializationProcessorType]): class SerializationProcessorRegistry(Registry[SerializationProcessorType]):
@ -1197,20 +1196,6 @@ class SerializationProcessorRegistry(Registry[SerializationProcessorType]):
name = "serialization_processors" name = "serialization_processors"
def get_all_for_mode(
self, import_export_mode: BaserowImportExportMode
) -> List[InstanceSubClass]:
"""
Returns registrables from `get_all`, filtered down to only
those that have a matching import/export mode.
"""
return [
registrable
for registrable in super().get_all()
if registrable.import_export_mode == import_export_mode
]
# A default plugin and application registry is created here, this is the one that is # A default plugin and application registry is created here, this is the one that is
# used throughout the whole Baserow application. To add a new plugin or application use # used throughout the whole Baserow application. To add a new plugin or application use

View file

@ -18,7 +18,7 @@ from baserow.core.handler import CoreHandler
from baserow.core.jobs.handler import JobHandler from baserow.core.jobs.handler import JobHandler
from baserow.core.jobs.models import Job from baserow.core.jobs.models import Job
from baserow.core.models import Application, Snapshot, User, Workspace from baserow.core.models import Application, Snapshot, User, Workspace
from baserow.core.registries import application_type_registry from baserow.core.registries import ImportExportConfig, application_type_registry
from baserow.core.signals import application_created from baserow.core.signals import application_created
from baserow.core.snapshots.exceptions import ( from baserow.core.snapshots.exceptions import (
MaximumSnapshotsReached, MaximumSnapshotsReached,
@ -380,9 +380,12 @@ class SnapshotHandler:
) )
application_type = application_type_registry.get_by_model(application) application_type = application_type_registry.get_by_model(application)
snapshot_import_export_config = ImportExportConfig(
include_permission_data=True, reduce_disk_space_usage=True
)
try: try:
exported_application = application_type.export_serialized( exported_application = application_type.export_serialized(
application, None, default_storage application, snapshot_import_export_config, None, default_storage
) )
except OperationalError as e: except OperationalError as e:
# Detect if this `OperationalError` is due to us exceeding the # Detect if this `OperationalError` is due to us exceeding the
@ -402,6 +405,7 @@ class SnapshotHandler:
application_type.import_serialized( application_type.import_serialized(
None, None,
exported_application, exported_application,
snapshot_import_export_config,
id_mapping, id_mapping,
None, None,
default_storage, default_storage,
@ -434,14 +438,19 @@ class SnapshotHandler:
application = snapshot.snapshot_to_application.specific application = snapshot.snapshot_to_application.specific
application_type = application_type_registry.get_by_model(application) application_type = application_type_registry.get_by_model(application)
restore_snapshot_import_export_config = ImportExportConfig(
include_permission_data=True, reduce_disk_space_usage=False
)
exported_application = application_type.export_serialized( exported_application = application_type.export_serialized(
application, None, default_storage application, restore_snapshot_import_export_config, None, default_storage
) )
progress.increment(by=50) progress.increment(by=50)
imported_application = application_type.import_serialized( imported_application = application_type.import_serialized(
snapshot.snapshot_from_application.workspace, snapshot.snapshot_from_application.workspace,
exported_application, exported_application,
restore_snapshot_import_export_config,
{}, {},
None, None,
default_storage, default_storage,

View file

@ -5,6 +5,7 @@ from baserow.contrib.builder.elements.models import HeadingElement, ParagraphEle
from baserow.contrib.builder.models import Builder from baserow.contrib.builder.models import Builder
from baserow.contrib.builder.pages.models import Page from baserow.contrib.builder.pages.models import Page
from baserow.core.db import specific_iterator from baserow.core.db import specific_iterator
from baserow.core.registries import ImportExportConfig
from baserow.core.trash.handler import TrashHandler from baserow.core.trash.handler import TrashHandler
@ -34,7 +35,9 @@ def test_builder_application_export(data_fixture):
element2 = data_fixture.create_builder_paragraph_element(page=page1) element2 = data_fixture.create_builder_paragraph_element(page=page1)
element3 = data_fixture.create_builder_heading_element(page=page2) element3 = data_fixture.create_builder_heading_element(page=page2)
serialized = BuilderApplicationType().export_serialized(builder) serialized = BuilderApplicationType().export_serialized(
builder, ImportExportConfig(include_permission_data=True)
)
assert serialized == { assert serialized == {
"pages": [ "pages": [
@ -143,8 +146,9 @@ def test_builder_application_import(data_fixture):
user = data_fixture.create_user() user = data_fixture.create_user()
workspace = data_fixture.create_workspace(user=user) workspace = data_fixture.create_workspace(user=user)
config = ImportExportConfig(include_permission_data=True)
builder = BuilderApplicationType().import_serialized( builder = BuilderApplicationType().import_serialized(
workspace, IMPORT_REFERENCE, {} workspace, IMPORT_REFERENCE, config, {}
) )
assert builder.id != IMPORT_REFERENCE["id"] assert builder.id != IMPORT_REFERENCE["id"]

View file

@ -10,6 +10,7 @@ from baserow.contrib.database.fields.handler import FieldHandler
from baserow.contrib.database.formula import BaserowFormulaNumberType from baserow.contrib.database.formula import BaserowFormulaNumberType
from baserow.contrib.database.rows.handler import RowHandler from baserow.contrib.database.rows.handler import RowHandler
from baserow.core.handler import CoreHandler from baserow.core.handler import CoreHandler
from baserow.core.registries import ImportExportConfig
@pytest.mark.django_db @pytest.mark.django_db
@ -367,11 +368,16 @@ def test_import_export_tables_with_count_fields(
through_field_id=link_row_field.id, through_field_id=link_row_field.id,
) )
config = ImportExportConfig(include_permission_data=False)
exported_applications = core_handler.export_workspace_applications( exported_applications = core_handler.export_workspace_applications(
database.workspace, BytesIO() database.workspace, BytesIO(), config
) )
imported_applications, id_mapping = core_handler.import_applications_to_workspace( imported_applications, id_mapping = core_handler.import_applications_to_workspace(
imported_workspace, exported_applications, BytesIO(), None imported_workspace,
exported_applications,
BytesIO(),
config,
None,
) )
imported_database = imported_applications[0] imported_database = imported_applications[0]
imported_tables = imported_database.table_set.all() imported_tables = imported_database.table_set.all()

View file

@ -11,6 +11,7 @@ from baserow.contrib.database.fields.handler import FieldHandler
from baserow.contrib.database.fields.models import CreatedOnField from baserow.contrib.database.fields.models import CreatedOnField
from baserow.contrib.database.rows.handler import RowHandler from baserow.contrib.database.rows.handler import RowHandler
from baserow.core.handler import CoreHandler from baserow.core.handler import CoreHandler
from baserow.core.registries import ImportExportConfig
@pytest.mark.django_db @pytest.mark.django_db
@ -182,8 +183,9 @@ def test_import_export_last_modified_field(data_fixture):
) )
core_handler = CoreHandler() core_handler = CoreHandler()
config = ImportExportConfig(include_permission_data=False)
exported_applications = core_handler.export_workspace_applications( exported_applications = core_handler.export_workspace_applications(
database.workspace, BytesIO() database.workspace, BytesIO(), config
) )
# We manually set this value in the export, because if it's set, then the import # We manually set this value in the export, because if it's set, then the import
@ -197,7 +199,7 @@ def test_import_export_last_modified_field(data_fixture):
imported_applications, imported_applications,
id_mapping, id_mapping,
) = core_handler.import_applications_to_workspace( ) = core_handler.import_applications_to_workspace(
imported_workspace, exported_applications, BytesIO(), None imported_workspace, exported_applications, BytesIO(), config, None
) )
imported_database = imported_applications[0] imported_database = imported_applications[0]

View file

@ -14,6 +14,7 @@ from baserow.contrib.database.fields.handler import FieldHandler
from baserow.contrib.database.fields.models import DateField from baserow.contrib.database.fields.models import DateField
from baserow.contrib.database.fields.registries import field_type_registry from baserow.contrib.database.fields.registries import field_type_registry
from baserow.contrib.database.rows.handler import RowHandler from baserow.contrib.database.rows.handler import RowHandler
from baserow.core.registries import ImportExportConfig
@pytest.mark.django_db @pytest.mark.django_db
@ -542,7 +543,11 @@ def test_import_export_date_field(data_fixture):
date_field_type = field_type_registry.get_by_model(date_field) date_field_type = field_type_registry.get_by_model(date_field)
number_serialized = date_field_type.export_serialized(date_field) number_serialized = date_field_type.export_serialized(date_field)
number_field_imported = date_field_type.import_serialized( number_field_imported = date_field_type.import_serialized(
date_field.table, number_serialized, {}, DeferredFieldFkUpdater() date_field.table,
number_serialized,
ImportExportConfig(include_permission_data=True),
{},
DeferredFieldFkUpdater(),
) )
assert date_field.date_format == number_field_imported.date_format assert date_field.date_format == number_field_imported.date_format
assert date_field.date_include_time == number_field_imported.date_include_time assert date_field.date_include_time == number_field_imported.date_include_time

View file

@ -17,6 +17,7 @@ from baserow.contrib.database.fields.models import (
) )
from baserow.contrib.database.fields.registries import FieldType, field_type_registry from baserow.contrib.database.fields.registries import FieldType, field_type_registry
from baserow.contrib.database.rows.handler import RowHandler from baserow.contrib.database.rows.handler import RowHandler
from baserow.core.registries import ImportExportConfig
from baserow.test_utils.helpers import setup_interesting_test_table from baserow.test_utils.helpers import setup_interesting_test_table
@ -30,7 +31,11 @@ def test_import_export_text_field(data_fixture):
text_field_type = field_type_registry.get_by_model(text_field) text_field_type = field_type_registry.get_by_model(text_field)
text_serialized = text_field_type.export_serialized(text_field) text_serialized = text_field_type.export_serialized(text_field)
text_field_imported = text_field_type.import_serialized( text_field_imported = text_field_type.import_serialized(
text_field.table, text_serialized, id_mapping, DeferredFieldFkUpdater() text_field.table,
text_serialized,
ImportExportConfig(include_permission_data=True),
id_mapping,
DeferredFieldFkUpdater(),
) )
assert text_field.id != text_field_imported.id assert text_field.id != text_field_imported.id
assert text_field.name == text_field_imported.name assert text_field.name == text_field_imported.name
@ -66,6 +71,7 @@ def test_import_export_formula_field(data_fixture, api_client):
formula_field_imported = formula_field_type.import_serialized( formula_field_imported = formula_field_type.import_serialized(
text_field_in_diff_table.table, text_field_in_diff_table.table,
formula_serialized, formula_serialized,
ImportExportConfig(include_permission_data=True),
id_mapping, id_mapping,
DeferredFieldFkUpdater(), DeferredFieldFkUpdater(),
) )
@ -670,7 +676,11 @@ def test_import_export_lookup_field(data_fixture, api_client):
deferred_field_fk_updater = DeferredFieldFkUpdater() deferred_field_fk_updater = DeferredFieldFkUpdater()
lookup_field_imported = lookup_field_type.import_serialized( lookup_field_imported = lookup_field_type.import_serialized(
table_a, lookup_serialized, id_mapping, deferred_field_fk_updater table_a,
lookup_serialized,
ImportExportConfig(include_permission_data=True),
id_mapping,
deferred_field_fk_updater,
) )
assert lookup.id != lookup_field_imported.id assert lookup.id != lookup_field_imported.id
assert lookup_field_imported.name == "lookup" assert lookup_field_imported.name == "lookup"

View file

@ -12,6 +12,7 @@ from baserow.contrib.database.fields.handler import FieldHandler
from baserow.contrib.database.fields.models import FileField from baserow.contrib.database.fields.models import FileField
from baserow.contrib.database.rows.handler import RowHandler from baserow.contrib.database.rows.handler import RowHandler
from baserow.core.handler import CoreHandler from baserow.core.handler import CoreHandler
from baserow.core.registries import ImportExportConfig
from baserow.core.user_files.exceptions import ( from baserow.core.user_files.exceptions import (
InvalidUserFileNameError, InvalidUserFileNameError,
UserFileDoesNotExist, UserFileDoesNotExist,
@ -225,8 +226,13 @@ def test_import_export_file_field(data_fixture, tmpdir):
) )
files_buffer = BytesIO() files_buffer = BytesIO()
config = ImportExportConfig(include_permission_data=False)
exported_applications = core_handler.export_workspace_applications( exported_applications = core_handler.export_workspace_applications(
database.workspace, files_buffer=files_buffer, storage=storage database.workspace,
files_buffer=files_buffer,
storage=storage,
import_export_config=config,
) )
# We expect that the exported zip file contains the user file used in the created # We expect that the exported zip file contains the user file used in the created
@ -259,7 +265,7 @@ def test_import_export_file_field(data_fixture, tmpdir):
] = "test2.txt" ] = "test2.txt"
imported_applications, id_mapping = core_handler.import_applications_to_workspace( imported_applications, id_mapping = core_handler.import_applications_to_workspace(
imported_workspace, exported_applications, files_buffer, storage imported_workspace, exported_applications, files_buffer, config, storage
) )
imported_database = imported_applications[0] imported_database = imported_applications[0]
imported_tables = imported_database.table_set.all() imported_tables = imported_database.table_set.all()

View file

@ -11,6 +11,7 @@ from baserow.contrib.database.fields.handler import FieldHandler
from baserow.contrib.database.fields.models import LastModifiedField from baserow.contrib.database.fields.models import LastModifiedField
from baserow.contrib.database.rows.handler import RowHandler from baserow.contrib.database.rows.handler import RowHandler
from baserow.core.handler import CoreHandler from baserow.core.handler import CoreHandler
from baserow.core.registries import ImportExportConfig
@pytest.mark.django_db @pytest.mark.django_db
@ -186,8 +187,10 @@ def test_import_export_last_modified_field(data_fixture):
) )
core_handler = CoreHandler() core_handler = CoreHandler()
config = ImportExportConfig(include_permission_data=False)
exported_applications = core_handler.export_workspace_applications( exported_applications = core_handler.export_workspace_applications(
database.workspace, BytesIO() database.workspace, BytesIO(), config
) )
# We manually set this value in the export, because if it's set, then the import # We manually set this value in the export, because if it's set, then the import
@ -201,7 +204,7 @@ def test_import_export_last_modified_field(data_fixture):
imported_applications, imported_applications,
id_mapping, id_mapping,
) = core_handler.import_applications_to_workspace( ) = core_handler.import_applications_to_workspace(
imported_workspace, exported_applications, BytesIO(), None imported_workspace, exported_applications, BytesIO(), config, None
) )
imported_database = imported_applications[0] imported_database = imported_applications[0]

View file

@ -23,6 +23,7 @@ from baserow.contrib.database.rows.handler import RowHandler
from baserow.contrib.database.table.handler import TableHandler from baserow.contrib.database.table.handler import TableHandler
from baserow.core.handler import CoreHandler from baserow.core.handler import CoreHandler
from baserow.core.models import TrashEntry from baserow.core.models import TrashEntry
from baserow.core.registries import ImportExportConfig
from baserow.core.trash.handler import TrashHandler from baserow.core.trash.handler import TrashHandler
@ -881,11 +882,13 @@ def test_import_export_link_row_field(data_fixture):
values={f"field_{link_row_field.id}": [c_row.id, c_row_2.id]}, values={f"field_{link_row_field.id}": [c_row.id, c_row_2.id]},
) )
config = ImportExportConfig(include_permission_data=False)
exported_applications = core_handler.export_workspace_applications( exported_applications = core_handler.export_workspace_applications(
database.workspace, BytesIO() database.workspace, BytesIO(), config
) )
imported_applications, id_mapping = core_handler.import_applications_to_workspace( imported_applications, id_mapping = core_handler.import_applications_to_workspace(
imported_workspace, exported_applications, BytesIO(), None imported_workspace, exported_applications, BytesIO(), config, None
) )
imported_database = imported_applications[0] imported_database = imported_applications[0]
imported_tables = imported_database.table_set.all() imported_tables = imported_database.table_set.all()

View file

@ -23,6 +23,7 @@ from baserow.contrib.database.rows.handler import RowHandler
from baserow.contrib.database.views.handler import ViewHandler from baserow.contrib.database.views.handler import ViewHandler
from baserow.core.db import specific_iterator from baserow.core.db import specific_iterator
from baserow.core.handler import CoreHandler from baserow.core.handler import CoreHandler
from baserow.core.registries import ImportExportConfig
@pytest.mark.django_db @pytest.mark.django_db
@ -457,7 +458,11 @@ def test_import_export_lookup_field_when_through_field_trashed(
lookup.save() lookup.save()
lookup_field_imported = lookup_field_type.import_serialized( lookup_field_imported = lookup_field_type.import_serialized(
table_a, lookup_serialized, id_mapping, DeferredFieldFkUpdater() table_a,
lookup_serialized,
ImportExportConfig(include_permission_data=True),
id_mapping,
DeferredFieldFkUpdater(),
) )
assert lookup_field_imported.through_field is None assert lookup_field_imported.through_field is None
assert lookup_field_imported.through_field_name == link_field.name assert lookup_field_imported.through_field_name == link_field.name
@ -508,7 +513,11 @@ def test_import_export_lookup_field_trashed_target_field(data_fixture, api_clien
lookup.save() lookup.save()
lookup_field_imported = lookup_field_type.import_serialized( lookup_field_imported = lookup_field_type.import_serialized(
table_a, lookup_serialized, id_mapping, DeferredFieldFkUpdater() table_a,
lookup_serialized,
ImportExportConfig(include_permission_data=True),
id_mapping,
DeferredFieldFkUpdater(),
) )
assert lookup_field_imported.through_field is None assert lookup_field_imported.through_field is None
assert lookup_field_imported.through_field_name == link_field.name assert lookup_field_imported.through_field_name == link_field.name
@ -578,11 +587,12 @@ def test_import_export_tables_with_lookup_fields(
target_field_id=customer_age.id, target_field_id=customer_age.id,
) )
config = ImportExportConfig(include_permission_data=False)
exported_applications = core_handler.export_workspace_applications( exported_applications = core_handler.export_workspace_applications(
database.workspace, BytesIO() database.workspace, BytesIO(), config
) )
imported_applications, id_mapping = core_handler.import_applications_to_workspace( imported_applications, id_mapping = core_handler.import_applications_to_workspace(
imported_workspace, exported_applications, BytesIO(), None imported_workspace, exported_applications, BytesIO(), config, None
) )
imported_database = imported_applications[0] imported_database = imported_applications[0]
imported_tables = imported_database.table_set.all() imported_tables = imported_database.table_set.all()

View file

@ -11,6 +11,7 @@ from baserow.contrib.database.fields.models import MultipleCollaboratorsField
from baserow.contrib.database.rows.handler import RowHandler from baserow.contrib.database.rows.handler import RowHandler
from baserow.contrib.database.views.handler import ViewHandler from baserow.contrib.database.views.handler import ViewHandler
from baserow.core.handler import CoreHandler from baserow.core.handler import CoreHandler
from baserow.core.registries import ImportExportConfig
@pytest.mark.django_db @pytest.mark.django_db
@ -131,11 +132,12 @@ def test_get_set_export_serialized_value_multiple_collaborators_field(data_fixtu
}, },
) )
config = ImportExportConfig(include_permission_data=False)
exported_applications = core_handler.export_workspace_applications( exported_applications = core_handler.export_workspace_applications(
workspace, BytesIO() workspace, BytesIO(), config
) )
imported_applications, id_mapping = core_handler.import_applications_to_workspace( imported_applications, id_mapping = core_handler.import_applications_to_workspace(
imported_workspace, exported_applications, BytesIO(), None imported_workspace, exported_applications, BytesIO(), config, None
) )
imported_database = imported_applications[0] imported_database = imported_applications[0]
imported_table = imported_database.table_set.all()[0] imported_table = imported_database.table_set.all()[0]

View file

@ -29,6 +29,7 @@ from baserow.contrib.database.fields.registries import field_type_registry
from baserow.contrib.database.rows.handler import RowHandler from baserow.contrib.database.rows.handler import RowHandler
from baserow.contrib.database.views.handler import ViewHandler from baserow.contrib.database.views.handler import ViewHandler
from baserow.core.handler import CoreHandler from baserow.core.handler import CoreHandler
from baserow.core.registries import ImportExportConfig
@pytest.mark.django_db @pytest.mark.django_db
@ -732,7 +733,11 @@ def test_import_export_multiple_select_field(data_fixture):
field_serialized = field_type.export_serialized(field) field_serialized = field_type.export_serialized(field)
id_mapping = {} id_mapping = {}
field_imported = field_type.import_serialized( field_imported = field_type.import_serialized(
table, field_serialized, id_mapping, DeferredFieldFkUpdater() table,
field_serialized,
ImportExportConfig(include_permission_data=True),
id_mapping,
DeferredFieldFkUpdater(),
) )
assert field_imported.select_options.all().count() == 4 assert field_imported.select_options.all().count() == 4
@ -798,11 +803,13 @@ def test_get_set_export_serialized_value_multiple_select_field(
) )
assert len(SelectOption.objects.all()) == 3 assert len(SelectOption.objects.all()) == 3
config = ImportExportConfig(include_permission_data=False)
exported_applications = core_handler.export_workspace_applications( exported_applications = core_handler.export_workspace_applications(
workspace, BytesIO() workspace, BytesIO(), config
) )
imported_applications, id_mapping = core_handler.import_applications_to_workspace( imported_applications, id_mapping = core_handler.import_applications_to_workspace(
imported_workspace, exported_applications, BytesIO(), None imported_workspace, exported_applications, BytesIO(), config, None
) )
imported_database = imported_applications[0] imported_database = imported_applications[0]
imported_table = imported_database.table_set.all()[0] imported_table = imported_database.table_set.all()[0]

View file

@ -11,6 +11,7 @@ from baserow.contrib.database.fields.handler import FieldHandler
from baserow.contrib.database.fields.models import NumberField from baserow.contrib.database.fields.models import NumberField
from baserow.contrib.database.fields.registries import field_type_registry from baserow.contrib.database.fields.registries import field_type_registry
from baserow.contrib.database.rows.handler import RowHandler from baserow.contrib.database.rows.handler import RowHandler
from baserow.core.registries import ImportExportConfig
@pytest.mark.django_db @pytest.mark.django_db
@ -198,7 +199,11 @@ def test_import_export_number_field(data_fixture):
number_field_type = field_type_registry.get_by_model(number_field) number_field_type = field_type_registry.get_by_model(number_field)
number_serialized = number_field_type.export_serialized(number_field) number_serialized = number_field_type.export_serialized(number_field)
number_field_imported = number_field_type.import_serialized( number_field_imported = number_field_type.import_serialized(
number_field.table, number_serialized, {}, DeferredFieldFkUpdater() number_field.table,
number_serialized,
ImportExportConfig(include_permission_data=True),
{},
DeferredFieldFkUpdater(),
) )
assert number_field.number_negative == number_field_imported.number_negative assert number_field.number_negative == number_field_imported.number_negative
assert number_field.number_decimal_places == ( assert number_field.number_decimal_places == (

View file

@ -14,6 +14,7 @@ from baserow.contrib.database.formula import BaserowFormulaNumberType
from baserow.contrib.database.formula.types.exceptions import InvalidFormulaType from baserow.contrib.database.formula.types.exceptions import InvalidFormulaType
from baserow.contrib.database.rows.handler import RowHandler from baserow.contrib.database.rows.handler import RowHandler
from baserow.core.handler import CoreHandler from baserow.core.handler import CoreHandler
from baserow.core.registries import ImportExportConfig
from baserow.formula.exceptions import FormulaFunctionTypeDoesNotExist from baserow.formula.exceptions import FormulaFunctionTypeDoesNotExist
@ -584,11 +585,12 @@ def test_import_export_tables_with_rollup_fields(
rollup_function="sum", rollup_function="sum",
) )
config = ImportExportConfig(include_permission_data=False)
exported_applications = core_handler.export_workspace_applications( exported_applications = core_handler.export_workspace_applications(
database.workspace, BytesIO() database.workspace, BytesIO(), config
) )
imported_applications, id_mapping = core_handler.import_applications_to_workspace( imported_applications, id_mapping = core_handler.import_applications_to_workspace(
imported_workspace, exported_applications, BytesIO(), None imported_workspace, exported_applications, BytesIO(), config, None
) )
imported_database = imported_applications[0] imported_database = imported_applications[0]
imported_tables = imported_database.table_set.all() imported_tables = imported_database.table_set.all()

View file

@ -21,6 +21,7 @@ from baserow.contrib.database.fields.registries import field_type_registry
from baserow.contrib.database.rows.handler import RowHandler from baserow.contrib.database.rows.handler import RowHandler
from baserow.contrib.database.views.handler import ViewHandler from baserow.contrib.database.views.handler import ViewHandler
from baserow.core.handler import CoreHandler from baserow.core.handler import CoreHandler
from baserow.core.registries import ImportExportConfig
@pytest.mark.django_db @pytest.mark.django_db
@ -950,7 +951,11 @@ def test_import_export_single_select_field(data_fixture):
field_serialized = field_type.export_serialized(field) field_serialized = field_type.export_serialized(field)
id_mapping = {} id_mapping = {}
field_imported = field_type.import_serialized( field_imported = field_type.import_serialized(
table, field_serialized, id_mapping, DeferredFieldFkUpdater() table,
field_serialized,
ImportExportConfig(include_permission_data=True),
id_mapping,
DeferredFieldFkUpdater(),
) )
assert field_imported.select_options.all().count() == 1 assert field_imported.select_options.all().count() == 1
@ -979,11 +984,12 @@ def test_get_set_export_serialized_value_single_select_field(data_fixture):
model.objects.create(**{f"field_{field.id}_id": option_a.id}) model.objects.create(**{f"field_{field.id}_id": option_a.id})
model.objects.create(**{f"field_{field.id}_id": option_b.id}) model.objects.create(**{f"field_{field.id}_id": option_b.id})
config = ImportExportConfig(include_permission_data=False)
exported_applications = core_handler.export_workspace_applications( exported_applications = core_handler.export_workspace_applications(
workspace, BytesIO() workspace, BytesIO(), config
) )
imported_applications, id_mapping = core_handler.import_applications_to_workspace( imported_applications, id_mapping = core_handler.import_applications_to_workspace(
imported_workspace, exported_applications, BytesIO(), None imported_workspace, exported_applications, BytesIO(), config, None
) )
imported_database = imported_applications[0] imported_database = imported_applications[0]
imported_table = imported_database.table_set.all()[0] imported_table = imported_database.table_set.all()[0]
@ -1028,11 +1034,13 @@ def test_get_set_export_serialized_value_single_select_field_with_deleted_option
# Deleting the option doesn't set the row value to None. # Deleting the option doesn't set the row value to None.
option_a.delete() option_a.delete()
config = ImportExportConfig(include_permission_data=False)
exported_applications = core_handler.export_workspace_applications( exported_applications = core_handler.export_workspace_applications(
workspace, BytesIO() workspace, BytesIO(), config
) )
imported_applications, id_mapping = core_handler.import_applications_to_workspace( imported_applications, id_mapping = core_handler.import_applications_to_workspace(
imported_workspace, exported_applications, BytesIO(), None imported_workspace, exported_applications, BytesIO(), config, None
) )
imported_database = imported_applications[0] imported_database = imported_applications[0]
imported_table = imported_database.table_set.all()[0] imported_table = imported_database.table_set.all()[0]

View file

@ -7,7 +7,7 @@ from pytz import UTC
from baserow.contrib.database.fields.models import FormulaField, TextField from baserow.contrib.database.fields.models import FormulaField, TextField
from baserow.contrib.database.table.models import Table from baserow.contrib.database.table.models import Table
from baserow.core.handler import CoreHandler from baserow.core.handler import CoreHandler
from baserow.core.registries import application_type_registry from baserow.core.registries import ImportExportConfig, application_type_registry
@pytest.mark.django_db @pytest.mark.django_db
@ -35,7 +35,8 @@ def test_import_export_database(data_fixture):
row.refresh_from_db() row.refresh_from_db()
database_type = application_type_registry.get("database") database_type = application_type_registry.get("database")
serialized = database_type.export_serialized(database, None, None) config = ImportExportConfig(include_permission_data=True)
serialized = database_type.export_serialized(database, config)
# Delete the updated on, because the import should also be compatible with # Delete the updated on, because the import should also be compatible with
# without these values present. # without these values present.
@ -47,7 +48,12 @@ def test_import_export_database(data_fixture):
with freeze_time("2022-01-01 12:00"): with freeze_time("2022-01-01 12:00"):
imported_database = database_type.import_serialized( imported_database = database_type.import_serialized(
imported_workspace, serialized, id_mapping, None, None imported_workspace,
serialized,
config,
id_mapping,
None,
None,
) )
assert imported_database.id != database.id assert imported_database.id != database.id

View file

@ -43,6 +43,7 @@ from baserow.core.models import (
WorkspaceUser, WorkspaceUser,
) )
from baserow.core.operations import ReadWorkspaceOperationType from baserow.core.operations import ReadWorkspaceOperationType
from baserow.core.registries import ImportExportConfig
from baserow.core.trash.handler import TrashHandler from baserow.core.trash.handler import TrashHandler
from baserow.core.user_files.models import UserFile from baserow.core.user_files.models import UserFile
@ -1141,9 +1142,12 @@ def test_export_import_workspace_application(data_fixture):
data_fixture.create_database_table(database=database) data_fixture.create_database_table(database=database)
handler = CoreHandler() handler = CoreHandler()
exported_applications = handler.export_workspace_applications(workspace, BytesIO()) config = ImportExportConfig(include_permission_data=False)
exported_applications = handler.export_workspace_applications(
workspace, BytesIO(), config
)
imported_applications, id_mapping = handler.import_applications_to_workspace( imported_applications, id_mapping = handler.import_applications_to_workspace(
imported_workspace, exported_applications, BytesIO(), None imported_workspace, exported_applications, BytesIO(), config, None
) )
assert len(imported_applications) == 1 assert len(imported_applications) == 1

View file

@ -9,11 +9,6 @@ from baserow.core.exceptions import (
InstanceTypeAlreadyRegistered, InstanceTypeAlreadyRegistered,
InstanceTypeDoesNotExist, InstanceTypeDoesNotExist,
) )
from baserow.core.registries import (
BaserowImportExportMode,
SerializationProcessorRegistry,
SerializationProcessorType,
)
from baserow.core.registry import ( from baserow.core.registry import (
CustomFieldsInstanceMixin, CustomFieldsInstanceMixin,
CustomFieldsRegistryMixin, CustomFieldsRegistryMixin,
@ -204,34 +199,3 @@ def test_get_serializer(data_fixture):
serializer = registry.get_serializer(database, request=True) serializer = registry.get_serializer(database, request=True)
assert "order" in serializer.data assert "order" in serializer.data
def test_serialization_processor_registry_get_all_for_mode_matching_mode():
class MatchingProcessorType(SerializationProcessorType):
type = "matching_processor_type"
import_export_mode = BaserowImportExportMode.TARGETING_SAME_WORKSPACE_NEW_PK
processor_type = MatchingProcessorType()
registry = SerializationProcessorRegistry()
registry.register(processor_type)
assert registry.get_all_for_mode(
BaserowImportExportMode.TARGETING_SAME_WORKSPACE_NEW_PK
) == [processor_type]
def test_serialization_processor_registry_get_all_for_mode_mismatching_mode():
class MatchingProcessorType(SerializationProcessorType):
type = "mismatching_processor_type"
import_export_mode = BaserowImportExportMode.TARGETING_SAME_WORKSPACE_NEW_PK
processor_type = MatchingProcessorType()
registry = SerializationProcessorRegistry()
registry.register(processor_type)
assert (
registry.get_all_for_mode(
BaserowImportExportMode.TARGETING_DIFF_WORKSPACE_NEW_PK
)
== []
)

View file

@ -5,7 +5,7 @@ from django.contrib.contenttypes.models import ContentType
from baserow_premium.license.handler import LicenseHandler from baserow_premium.license.handler import LicenseHandler
from baserow.core.models import Application from baserow.core.models import Application
from baserow.core.registries import BaserowImportExportMode, SerializationProcessorType from baserow.core.registries import ImportExportConfig, SerializationProcessorType
from baserow.core.types import SerializationProcessorScope from baserow.core.types import SerializationProcessorScope
from baserow.core.utils import atomic_if_not_already from baserow.core.utils import atomic_if_not_already
from baserow_enterprise.features import RBAC from baserow_enterprise.features import RBAC
@ -30,7 +30,6 @@ class EnterpriseExportSerializedStructure:
class RoleAssignmentSerializationProcessorType(SerializationProcessorType): class RoleAssignmentSerializationProcessorType(SerializationProcessorType):
type = "role_assignment_serialization_processors" type = "role_assignment_serialization_processors"
structure = EnterpriseExportSerializedStructure structure = EnterpriseExportSerializedStructure
import_export_mode = BaserowImportExportMode.TARGETING_SAME_WORKSPACE_NEW_PK
@classmethod @classmethod
def import_serialized( def import_serialized(
@ -38,12 +37,19 @@ class RoleAssignmentSerializationProcessorType(SerializationProcessorType):
workspace: "Workspace", workspace: "Workspace",
scope: SerializationProcessorScope, scope: SerializationProcessorScope,
serialized_scope: dict, serialized_scope: dict,
) -> None: import_export_config: ImportExportConfig,
):
""" """
Responsible for importing any `role_assignments` in `serialized_scope` Responsible for importing any `role_assignments` in `serialized_scope`
into a newly restored/duplicated scope in `scope`. into a newly restored/duplicated scope in `scope`.
""" """
if not import_export_config.include_permission_data:
# We cannot yet export RBAC roles to another workspace as we would also need
# to export all subjects to the new workspace also or somehow allow to user
# to choose how to map subjects.
return
# Application subclass scopes can't be passed to # Application subclass scopes can't be passed to
# the role assignment handler. See #1624. # the role assignment handler. See #1624.
if isinstance(scope, Application): if isinstance(scope, Application):
@ -71,12 +77,19 @@ class RoleAssignmentSerializationProcessorType(SerializationProcessorType):
cls, cls,
workspace: "Workspace", workspace: "Workspace",
scope: SerializationProcessorScope, scope: SerializationProcessorScope,
import_export_config: ImportExportConfig,
) -> dict[str, Any]: ) -> dict[str, Any]:
""" """
Exports the `role_assignments` in `scope` when it is being exported Exports the `role_assignments` in `scope` when it is being exported
by an application type `export_serialized`. by an application type `export_serialized`.
""" """
if not import_export_config.include_permission_data:
# We cannot yet export RBAC roles to another workspace as we would also need
# to export all subjects to the new workspace also or somehow allow to user
# to choose how to map subjects.
return
# Do not export anything if the workspace doesn't have RBAC enabled. # Do not export anything if the workspace doesn't have RBAC enabled.
if not LicenseHandler.workspace_has_feature(RBAC, workspace): if not LicenseHandler.workspace_has_feature(RBAC, workspace):
return {} return {}

View file

@ -3,6 +3,7 @@ from django.contrib.contenttypes.models import ContentType
import pytest import pytest
from baserow.contrib.database.table.handler import TableHandler from baserow.contrib.database.table.handler import TableHandler
from baserow.core.registries import ImportExportConfig
from baserow_enterprise.role.handler import RoleAssignmentHandler from baserow_enterprise.role.handler import RoleAssignmentHandler
from baserow_enterprise.role.models import Role from baserow_enterprise.role.models import Role
from baserow_enterprise.structure_types import RoleAssignmentSerializationProcessorType from baserow_enterprise.structure_types import RoleAssignmentSerializationProcessorType
@ -21,10 +22,12 @@ def test_export_serialized_structure_on_database(enterprise_data_fixture):
database = enterprise_data_fixture.create_database_application(workspace=workspace) database = enterprise_data_fixture.create_database_application(workspace=workspace)
application = database.application_ptr application = database.application_ptr
config = ImportExportConfig(include_permission_data=True)
role = Role.objects.get(uid="ADMIN") role = Role.objects.get(uid="ADMIN")
RoleAssignmentHandler().assign_role(user, workspace, role, application) RoleAssignmentHandler().assign_role(user, workspace, role, application)
serialized_structure = enterprise_structure.export_serialized( serialized_structure = enterprise_structure.export_serialized(
workspace, application workspace, application, config
) )
content_types = ContentType.objects.get_for_models(user, application) content_types = ContentType.objects.get_for_models(user, application)
@ -49,8 +52,9 @@ def test_import_serialized_structure_on_database(enterprise_data_fixture):
role = Role.objects.get(uid="ADMIN") role = Role.objects.get(uid="ADMIN")
RoleAssignmentHandler().assign_role(user, workspace, role, application) RoleAssignmentHandler().assign_role(user, workspace, role, application)
config = ImportExportConfig(include_permission_data=True)
serialized_structure = enterprise_structure.export_serialized( serialized_structure = enterprise_structure.export_serialized(
workspace, application workspace, application, config
) )
new_database = enterprise_data_fixture.create_database_application( new_database = enterprise_data_fixture.create_database_application(
@ -59,7 +63,7 @@ def test_import_serialized_structure_on_database(enterprise_data_fixture):
new_application = new_database.application_ptr new_application = new_database.application_ptr
enterprise_structure.import_serialized( enterprise_structure.import_serialized(
workspace, new_application, serialized_structure workspace, new_application, serialized_structure, config
) )
role_assignments = RoleAssignmentHandler().get_role_assignments( role_assignments = RoleAssignmentHandler().get_role_assignments(
@ -82,11 +86,15 @@ def test_export_serialized_structure_on_table(enterprise_data_fixture):
workspace = enterprise_data_fixture.create_workspace(user=user) workspace = enterprise_data_fixture.create_workspace(user=user)
database = enterprise_data_fixture.create_database_application(workspace=workspace) database = enterprise_data_fixture.create_database_application(workspace=workspace)
config = ImportExportConfig(include_permission_data=True)
role = Role.objects.get(uid="ADMIN") role = Role.objects.get(uid="ADMIN")
table, _ = TableHandler().create_table(user, database, name="Table") table, _ = TableHandler().create_table(user, database, name="Table")
RoleAssignmentHandler().assign_role(user, workspace, role, table) RoleAssignmentHandler().assign_role(user, workspace, role, table)
serialized_structure = enterprise_structure.export_serialized(workspace, table) serialized_structure = enterprise_structure.export_serialized(
workspace, table, config
)
content_types = ContentType.objects.get_for_models(user, table) content_types = ContentType.objects.get_for_models(user, table)
assert serialized_structure == { assert serialized_structure == {
@ -107,13 +115,19 @@ def test_import_serialized_structure_on_table(enterprise_data_fixture):
workspace = enterprise_data_fixture.create_workspace(user=user) workspace = enterprise_data_fixture.create_workspace(user=user)
database = enterprise_data_fixture.create_database_application(workspace=workspace) database = enterprise_data_fixture.create_database_application(workspace=workspace)
config = ImportExportConfig(include_permission_data=True)
role = Role.objects.get(uid="ADMIN") role = Role.objects.get(uid="ADMIN")
table, _ = TableHandler().create_table(user, database, name="Table") table, _ = TableHandler().create_table(user, database, name="Table")
RoleAssignmentHandler().assign_role(user, workspace, role, table) RoleAssignmentHandler().assign_role(user, workspace, role, table)
serialized_structure = enterprise_structure.export_serialized(workspace, table) serialized_structure = enterprise_structure.export_serialized(
workspace, table, config
)
new_table, _ = TableHandler().create_table(user, database, name="New table") new_table, _ = TableHandler().create_table(user, database, name="New table")
enterprise_structure.import_serialized(workspace, new_table, serialized_structure) enterprise_structure.import_serialized(
workspace, new_table, serialized_structure, config
)
role_assignments = RoleAssignmentHandler().get_role_assignments( role_assignments = RoleAssignmentHandler().get_role_assignments(
workspace, new_table workspace, new_table