mirror of
https://github.com/p2p-ld/nwb-linkml.git
synced 2024-11-10 00:34:29 +00:00
sneakers-the-rat
fd9aef9531
- split off generated subclasses into "include" files, not sure if that's good, but in any case we have them separable now. - more work on cleanly un-nesting scalar and 1D-vector data into attributes and lists, respectively - brought the pydantic generator in-repo to do a bunch of overrides
119 lines
4.1 KiB
Python
119 lines
4.1 KiB
Python
"""
|
|
Namespaces adapter
|
|
|
|
Wraps the :class:`nwb_schema_language.Namespaces` and other objects with convenience methods
|
|
for extracting information and generating translated schema
|
|
"""
|
|
import pdb
|
|
from typing import List, Optional
|
|
from pydantic import BaseModel, Field, validator, PrivateAttr
|
|
from pprint import pformat
|
|
from linkml_runtime.linkml_model import SchemaDefinition
|
|
|
|
from nwb_schema_language import Namespaces
|
|
|
|
from nwb_linkml.adapters.adapter import Adapter, BuildResult
|
|
from nwb_linkml.adapters.schema import SchemaAdapter
|
|
from nwb_linkml.lang_elements import NwbLangSchema
|
|
|
|
class NamespacesAdapter(Adapter):
|
|
namespaces: Namespaces
|
|
schemas: List[SchemaAdapter]
|
|
imported: List['NamespacesAdapter'] = Field(default_factory=list)
|
|
|
|
_imports_populated = PrivateAttr(False)
|
|
|
|
def __init__(self, **kwargs):
|
|
super(NamespacesAdapter, self).__init__(**kwargs)
|
|
self._populate_schema_namespaces()
|
|
|
|
def build(self) -> BuildResult:
|
|
if not self._imports_populated:
|
|
self.populate_imports()
|
|
|
|
|
|
sch_result = BuildResult()
|
|
for sch in self.schemas:
|
|
sch_result += sch.build()
|
|
# recursive step
|
|
for imported in self.imported:
|
|
imported_build = imported.build()
|
|
sch_result += imported_build
|
|
|
|
# add in monkeypatch nwb types
|
|
sch_result.schemas.append(NwbLangSchema)
|
|
|
|
# now generate the top-level namespaces that import everything
|
|
for ns in self.namespaces.namespaces:
|
|
ns_schemas = [sch for sch in self.schemas if sch.namespace == ns.name]
|
|
ns_schema = SchemaDefinition(
|
|
name = ns.name,
|
|
id = ns.name,
|
|
description = ns.doc,
|
|
version = ns.version,
|
|
imports=[sch.name for sch in ns_schemas]
|
|
)
|
|
sch_result.schemas.append(ns_schema)
|
|
|
|
return sch_result
|
|
|
|
def _populate_schema_namespaces(self):
|
|
# annotate for each schema which namespace imports it
|
|
for sch in self.schemas:
|
|
# imports seem to always be from same folder, so we can just use name part
|
|
sch_name = sch.path.name
|
|
# find which namespace imports this schema file
|
|
for ns in self.namespaces.namespaces:
|
|
sources = [sch.source for sch in ns.schema_]
|
|
if sch_name in sources:
|
|
sch.namespace = ns.name
|
|
break
|
|
|
|
|
|
|
|
def find_type_source(self, name:str) -> SchemaAdapter:
|
|
"""
|
|
Given some neurodata_type_inc, find the schema that it's defined in.
|
|
"""
|
|
# First check within the main schema
|
|
internal_matches = []
|
|
for schema in self.schemas:
|
|
class_names = [cls.neurodata_type_def for cls in schema.created_classes]
|
|
if name in class_names:
|
|
internal_matches.append(schema)
|
|
|
|
import_matches = []
|
|
for imported_ns in self.imported:
|
|
for schema in imported_ns.schemas:
|
|
class_names = [cls.neurodata_type_def for cls in schema.created_classes]
|
|
if name in class_names:
|
|
import_matches.append(schema)
|
|
|
|
all_matches = [*internal_matches, *import_matches]
|
|
|
|
if len(all_matches)>1:
|
|
raise KeyError(f"Found multiple schemas in namespace that define {name}:\ninternal: {pformat(internal_matches)}\nimported:{pformat(import_matches)}")
|
|
elif len(all_matches) == 1:
|
|
return all_matches[0]
|
|
else:
|
|
raise KeyError(f"No schema found that define {name}")
|
|
|
|
def populate_imports(self):
|
|
"""
|
|
Populate the imports that are needed for each schema file
|
|
|
|
"""
|
|
for sch in self.schemas:
|
|
for needs in sch.needed_imports:
|
|
# shouldn't be recursive references, since imports should form a tree
|
|
depends_on = self.find_type_source(needs)
|
|
if depends_on not in sch.imports:
|
|
sch.imports.append(depends_on)
|
|
|
|
# do so recursively
|
|
for imported in self.imported:
|
|
imported.populate_imports()
|
|
|
|
self._imports_populated = True
|
|
|
|
|