@@ -32,8 +32,8 @@ class DataRowMetadataSchema(BaseModel):
32
32
33
33
@property
34
34
def id (self ):
35
- """ DataRowMetadataSchema.id is being deprecated after version 3.9
36
- in favor of DataRowMetadataSchema.uid
35
+ """ ` DataRowMetadataSchema.id is being deprecated after version 3.9
36
+ in favor of DataRowMetadataSchema.uid`
37
37
"""
38
38
warnings .warn ("`id` is being deprecated in favor of `uid`" )
39
39
return self .uid
@@ -46,7 +46,6 @@ def id(self):
46
46
47
47
48
48
class _CamelCaseMixin (BaseModel ):
49
-
50
49
class Config :
51
50
allow_population_by_field_name = True
52
51
alias_generator = camel_case
@@ -112,7 +111,6 @@ class DataRowMetadataOntology:
112
111
>>> mdo = client.get_data_row_metadata_ontology()
113
112
114
113
"""
115
-
116
114
def __init__ (self , client ):
117
115
118
116
self ._client = client
@@ -132,8 +130,8 @@ def _build_ontology(self):
132
130
]
133
131
self .reserved_by_id = self ._make_id_index (self .reserved_fields )
134
132
self .reserved_by_name : Dict [
135
- str ,
136
- DataRowMetadataSchema ] = self . _make_name_index ( self .reserved_fields )
133
+ str , DataRowMetadataSchema ] = self . _make_name_index (
134
+ self .reserved_fields )
137
135
138
136
# custom fields
139
137
self .custom_fields : List [DataRowMetadataSchema ] = [
@@ -209,9 +207,8 @@ def _parse_ontology(raw_ontology) -> List[DataRowMetadataSchema]:
209
207
return fields
210
208
211
209
def parse_metadata (
212
- self , unparsed : List [Dict [str ,
213
- List [Union [str ,
214
- Dict ]]]]) -> List [DataRowMetadata ]:
210
+ self , unparsed : List [Dict [str , List [Union [str , Dict ]]]]
211
+ ) -> List [DataRowMetadata ]:
215
212
""" Parse metadata responses
216
213
217
214
>>> mdo.parse_metadata([metdata])
@@ -304,8 +301,8 @@ def _batch_upsert(
304
301
data_row_id = m .data_row_id ,
305
302
fields = list (
306
303
chain .from_iterable (
307
- self ._parse_upsert (m ) for m in m . fields ))). dict (
308
- by_alias = True ))
304
+ self ._parse_upsert (m )
305
+ for m in m . fields ))). dict ( by_alias = True ))
309
306
310
307
res = _batch_operations (_batch_upsert , items , self ._batch_size )
311
308
return res
@@ -471,11 +468,12 @@ def _batch_operations(
471
468
472
469
473
470
def _validate_parse_embedding (
474
- field : DataRowMetadataField
471
+ field : DataRowMetadataField
475
472
) -> List [Dict [str , Union [SchemaId , Embedding ]]]:
476
473
477
474
if isinstance (field .value , list ):
478
- if not (Embedding .min_items <= len (field .value ) <= Embedding .max_items ):
475
+ if not (Embedding .min_items <= len (field .value ) <=
476
+ Embedding .max_items ):
479
477
raise ValueError (
480
478
"Embedding length invalid. "
481
479
"Must have length within the interval "
0 commit comments